var/home/core/zuul-output/0000755000175000017500000000000015070121234014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070135345015474 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004754625515070135334017720 0ustar rootrootOct 04 04:34:07 crc systemd[1]: Starting Kubernetes Kubelet... Oct 04 04:34:07 crc restorecon[4553]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:07 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:34:08 crc restorecon[4553]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 04 04:34:09 crc kubenswrapper[4575]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 04:34:09 crc kubenswrapper[4575]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 04 04:34:09 crc kubenswrapper[4575]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 04:34:09 crc kubenswrapper[4575]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 04:34:09 crc kubenswrapper[4575]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 04 04:34:09 crc kubenswrapper[4575]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.080777 4575 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089024 4575 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089046 4575 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089052 4575 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089058 4575 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089063 4575 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089069 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089074 4575 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089079 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089085 4575 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089090 4575 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089095 4575 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089102 4575 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089106 4575 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089112 4575 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089117 4575 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089121 4575 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089126 4575 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089131 4575 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089136 4575 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089140 4575 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089146 4575 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089150 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089155 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089162 4575 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089169 4575 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089175 4575 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089180 4575 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089189 4575 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089194 4575 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089200 4575 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089205 4575 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089210 4575 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089215 4575 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089219 4575 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089224 4575 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089229 4575 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089234 4575 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089239 4575 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089245 4575 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089250 4575 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089255 4575 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089260 4575 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089264 4575 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089269 4575 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089274 4575 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089278 4575 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089285 4575 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089291 4575 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089295 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089300 4575 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089305 4575 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089311 4575 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089317 4575 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089324 4575 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089330 4575 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089335 4575 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089340 4575 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089346 4575 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089351 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089356 4575 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089364 4575 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089370 4575 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089376 4575 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089383 4575 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089388 4575 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089394 4575 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089399 4575 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089404 4575 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089409 4575 feature_gate.go:330] unrecognized feature gate: Example Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089414 4575 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.089419 4575 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090292 4575 flags.go:64] FLAG: --address="0.0.0.0" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090309 4575 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090321 4575 flags.go:64] FLAG: --anonymous-auth="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090328 4575 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090335 4575 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090341 4575 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090348 4575 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090355 4575 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090361 4575 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090367 4575 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090373 4575 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090380 4575 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090387 4575 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090393 4575 flags.go:64] FLAG: --cgroup-root="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090398 4575 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090404 4575 flags.go:64] FLAG: --client-ca-file="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090410 4575 flags.go:64] FLAG: --cloud-config="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090416 4575 flags.go:64] FLAG: --cloud-provider="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090421 4575 flags.go:64] FLAG: --cluster-dns="[]" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090427 4575 flags.go:64] FLAG: --cluster-domain="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090433 4575 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090439 4575 flags.go:64] FLAG: --config-dir="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090444 4575 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090450 4575 flags.go:64] FLAG: --container-log-max-files="5" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090464 4575 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090470 4575 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090476 4575 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090481 4575 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090487 4575 flags.go:64] FLAG: --contention-profiling="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090493 4575 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090498 4575 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090505 4575 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090511 4575 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090518 4575 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090523 4575 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090529 4575 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090534 4575 flags.go:64] FLAG: --enable-load-reader="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090540 4575 flags.go:64] FLAG: --enable-server="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090547 4575 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090554 4575 flags.go:64] FLAG: --event-burst="100" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090561 4575 flags.go:64] FLAG: --event-qps="50" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090567 4575 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090573 4575 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090579 4575 flags.go:64] FLAG: --eviction-hard="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090608 4575 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090613 4575 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090619 4575 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090625 4575 flags.go:64] FLAG: --eviction-soft="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090632 4575 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090637 4575 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090643 4575 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090649 4575 flags.go:64] FLAG: --experimental-mounter-path="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090654 4575 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090660 4575 flags.go:64] FLAG: --fail-swap-on="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090665 4575 flags.go:64] FLAG: --feature-gates="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090672 4575 flags.go:64] FLAG: --file-check-frequency="20s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090678 4575 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090684 4575 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090689 4575 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090695 4575 flags.go:64] FLAG: --healthz-port="10248" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090701 4575 flags.go:64] FLAG: --help="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090707 4575 flags.go:64] FLAG: --hostname-override="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090712 4575 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090718 4575 flags.go:64] FLAG: --http-check-frequency="20s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090724 4575 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090730 4575 flags.go:64] FLAG: --image-credential-provider-config="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090735 4575 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090740 4575 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090746 4575 flags.go:64] FLAG: --image-service-endpoint="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090751 4575 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090757 4575 flags.go:64] FLAG: --kube-api-burst="100" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090762 4575 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090769 4575 flags.go:64] FLAG: --kube-api-qps="50" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090774 4575 flags.go:64] FLAG: --kube-reserved="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090780 4575 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090785 4575 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090791 4575 flags.go:64] FLAG: --kubelet-cgroups="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090797 4575 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090802 4575 flags.go:64] FLAG: --lock-file="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090808 4575 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090814 4575 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090819 4575 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090827 4575 flags.go:64] FLAG: --log-json-split-stream="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090835 4575 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090843 4575 flags.go:64] FLAG: --log-text-split-stream="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090850 4575 flags.go:64] FLAG: --logging-format="text" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090857 4575 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090865 4575 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090871 4575 flags.go:64] FLAG: --manifest-url="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090878 4575 flags.go:64] FLAG: --manifest-url-header="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090886 4575 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.090892 4575 flags.go:64] FLAG: --max-open-files="1000000" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091048 4575 flags.go:64] FLAG: --max-pods="110" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091055 4575 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091063 4575 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091070 4575 flags.go:64] FLAG: --memory-manager-policy="None" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091077 4575 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091084 4575 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091092 4575 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091099 4575 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091114 4575 flags.go:64] FLAG: --node-status-max-images="50" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091121 4575 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091128 4575 flags.go:64] FLAG: --oom-score-adj="-999" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091140 4575 flags.go:64] FLAG: --pod-cidr="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091147 4575 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091157 4575 flags.go:64] FLAG: --pod-manifest-path="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091162 4575 flags.go:64] FLAG: --pod-max-pids="-1" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091168 4575 flags.go:64] FLAG: --pods-per-core="0" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091174 4575 flags.go:64] FLAG: --port="10250" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091180 4575 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091186 4575 flags.go:64] FLAG: --provider-id="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091191 4575 flags.go:64] FLAG: --qos-reserved="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091197 4575 flags.go:64] FLAG: --read-only-port="10255" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091202 4575 flags.go:64] FLAG: --register-node="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091208 4575 flags.go:64] FLAG: --register-schedulable="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091213 4575 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091227 4575 flags.go:64] FLAG: --registry-burst="10" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091232 4575 flags.go:64] FLAG: --registry-qps="5" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091238 4575 flags.go:64] FLAG: --reserved-cpus="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091244 4575 flags.go:64] FLAG: --reserved-memory="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091254 4575 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091266 4575 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091276 4575 flags.go:64] FLAG: --rotate-certificates="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091286 4575 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091295 4575 flags.go:64] FLAG: --runonce="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091305 4575 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091315 4575 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091325 4575 flags.go:64] FLAG: --seccomp-default="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091335 4575 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091345 4575 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091355 4575 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091365 4575 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091375 4575 flags.go:64] FLAG: --storage-driver-password="root" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091385 4575 flags.go:64] FLAG: --storage-driver-secure="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091394 4575 flags.go:64] FLAG: --storage-driver-table="stats" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091406 4575 flags.go:64] FLAG: --storage-driver-user="root" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091416 4575 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091427 4575 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091437 4575 flags.go:64] FLAG: --system-cgroups="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091446 4575 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091461 4575 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091470 4575 flags.go:64] FLAG: --tls-cert-file="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091480 4575 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091491 4575 flags.go:64] FLAG: --tls-min-version="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091501 4575 flags.go:64] FLAG: --tls-private-key-file="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091510 4575 flags.go:64] FLAG: --topology-manager-policy="none" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091520 4575 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091530 4575 flags.go:64] FLAG: --topology-manager-scope="container" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091540 4575 flags.go:64] FLAG: --v="2" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091551 4575 flags.go:64] FLAG: --version="false" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091562 4575 flags.go:64] FLAG: --vmodule="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091573 4575 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.091614 4575 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091835 4575 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091846 4575 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091859 4575 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091867 4575 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091879 4575 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091890 4575 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091901 4575 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091911 4575 feature_gate.go:330] unrecognized feature gate: Example Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091920 4575 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091930 4575 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091939 4575 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091949 4575 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091958 4575 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091969 4575 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091983 4575 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.091992 4575 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092003 4575 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092012 4575 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092021 4575 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092030 4575 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092039 4575 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092047 4575 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092055 4575 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092064 4575 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092073 4575 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092083 4575 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092091 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092101 4575 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092109 4575 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092117 4575 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092125 4575 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092134 4575 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092142 4575 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092153 4575 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092164 4575 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092175 4575 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092184 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092194 4575 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092204 4575 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092213 4575 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092223 4575 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092232 4575 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092241 4575 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092249 4575 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092258 4575 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092266 4575 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092278 4575 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092286 4575 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092295 4575 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092304 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092312 4575 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092320 4575 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092329 4575 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092338 4575 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092346 4575 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092355 4575 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092363 4575 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092372 4575 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092380 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092389 4575 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092397 4575 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092405 4575 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092414 4575 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092422 4575 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092430 4575 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092439 4575 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092447 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092456 4575 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092471 4575 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092479 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.092487 4575 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.092501 4575 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.103939 4575 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.103985 4575 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104110 4575 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104122 4575 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104133 4575 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104141 4575 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104150 4575 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104159 4575 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104168 4575 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104176 4575 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104185 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104193 4575 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104201 4575 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104210 4575 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104219 4575 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104228 4575 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104236 4575 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104244 4575 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104253 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104262 4575 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104270 4575 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104279 4575 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104287 4575 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104296 4575 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104304 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104312 4575 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104320 4575 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104329 4575 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104337 4575 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104346 4575 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104354 4575 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104363 4575 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104372 4575 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104380 4575 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104389 4575 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104397 4575 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104408 4575 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104420 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104432 4575 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104442 4575 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104450 4575 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104458 4575 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104466 4575 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104478 4575 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104489 4575 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104498 4575 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104507 4575 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104515 4575 feature_gate.go:330] unrecognized feature gate: Example Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104524 4575 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104533 4575 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104541 4575 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104549 4575 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104558 4575 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104566 4575 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104577 4575 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104613 4575 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104622 4575 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104631 4575 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104640 4575 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104649 4575 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104657 4575 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104668 4575 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104679 4575 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104689 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104697 4575 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104706 4575 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104714 4575 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104725 4575 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104733 4575 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104741 4575 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104750 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104761 4575 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.104773 4575 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.104786 4575 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105011 4575 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105022 4575 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105032 4575 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105043 4575 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105055 4575 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105064 4575 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105073 4575 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105082 4575 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105091 4575 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105100 4575 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105108 4575 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105117 4575 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105126 4575 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105134 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105142 4575 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105151 4575 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105159 4575 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105168 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105176 4575 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105185 4575 feature_gate.go:330] unrecognized feature gate: Example Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105193 4575 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105201 4575 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105210 4575 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105218 4575 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105229 4575 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105240 4575 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105249 4575 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105258 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105267 4575 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105277 4575 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105285 4575 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105294 4575 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105304 4575 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105313 4575 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105323 4575 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105332 4575 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105341 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105350 4575 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105360 4575 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105368 4575 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105376 4575 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105385 4575 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105393 4575 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105402 4575 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105412 4575 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105423 4575 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105432 4575 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105441 4575 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105449 4575 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105458 4575 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105466 4575 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105475 4575 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105486 4575 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105495 4575 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105504 4575 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105513 4575 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105522 4575 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105530 4575 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105539 4575 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105548 4575 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105556 4575 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105564 4575 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105573 4575 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105581 4575 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105614 4575 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105622 4575 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105630 4575 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105638 4575 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105647 4575 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105655 4575 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.105665 4575 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.105676 4575 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.105919 4575 server.go:940] "Client rotation is on, will bootstrap in background" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.112325 4575 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.112464 4575 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.114162 4575 server.go:997] "Starting client certificate rotation" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.114220 4575 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.115317 4575 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-21 00:41:36.214511277 +0000 UTC Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.115399 4575 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1148h7m27.099114327s for next certificate rotation Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.143437 4575 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.147341 4575 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.166920 4575 log.go:25] "Validated CRI v1 runtime API" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.201707 4575 log.go:25] "Validated CRI v1 image API" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.203508 4575 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.211502 4575 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-04-04-27-42-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.211701 4575 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.223244 4575 manager.go:217] Machine: {Timestamp:2025-10-04 04:34:09.220530895 +0000 UTC m=+0.549089729 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1317216c-ca93-49ac-8b83-e45b26a19814 BootID:27d90f2a-0402-4831-8abb-70fda13695ec Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:1d:c4:b1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:1d:c4:b1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:1d:ed:00 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9a:4a:2d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:98:ca:fb Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:43:fe:60 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:56:52:80:cf:b7:66 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3a:f0:10:e5:57:ba Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.223412 4575 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.223621 4575 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.225893 4575 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.226149 4575 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.226194 4575 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.226443 4575 topology_manager.go:138] "Creating topology manager with none policy" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.226457 4575 container_manager_linux.go:303] "Creating device plugin manager" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.227207 4575 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.227231 4575 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.228457 4575 state_mem.go:36] "Initialized new in-memory state store" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.228857 4575 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.232383 4575 kubelet.go:418] "Attempting to sync node with API server" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.232402 4575 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.232433 4575 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.232446 4575 kubelet.go:324] "Adding apiserver pod source" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.232457 4575 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.238053 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.238220 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.238119 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.238430 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.239388 4575 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.241238 4575 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.242445 4575 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244144 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244167 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244174 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244180 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244191 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244199 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244206 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244216 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244223 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244229 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244242 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.244249 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.245247 4575 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.245628 4575 server.go:1280] "Started kubelet" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.246691 4575 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:09 crc systemd[1]: Started Kubernetes Kubelet. Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.250923 4575 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.250959 4575 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.252044 4575 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 05:38:57.811643699 +0000 UTC Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.252121 4575 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1993h4m48.559528222s for next certificate rotation Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.252475 4575 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.252502 4575 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.252736 4575 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.250369 4575 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.250342 4575 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.256023 4575 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.256818 4575 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.257950 4575 factory.go:55] Registering systemd factory Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.258881 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.259004 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.259098 4575 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.199:6443: connect: connection refused" interval="200ms" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.260943 4575 factory.go:221] Registration of the systemd container factory successfully Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.261120 4575 server.go:460] "Adding debug handlers to kubelet server" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.262434 4575 factory.go:153] Registering CRI-O factory Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.262483 4575 factory.go:221] Registration of the crio container factory successfully Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.262532 4575 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.262550 4575 factory.go:103] Registering Raw factory Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.262563 4575 manager.go:1196] Started watching for new ooms in manager Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.267368 4575 manager.go:319] Starting recovery of all containers Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270282 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270327 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270343 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270358 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270370 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270382 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270393 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270406 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270422 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270434 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270446 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270459 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270474 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270488 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270499 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270512 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270526 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270540 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270556 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270574 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270614 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270630 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270647 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270661 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270673 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270685 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270701 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270714 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270727 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270740 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270771 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270784 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270796 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270810 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270821 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270833 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270845 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270857 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.269305 4575 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.199:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b2fa14cf3ed7d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 04:34:09.245605245 +0000 UTC m=+0.574164049,LastTimestamp:2025-10-04 04:34:09.245605245 +0000 UTC m=+0.574164049,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270869 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270927 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270964 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270978 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.270989 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271001 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271147 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271162 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271175 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271187 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271199 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271211 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271222 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271234 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271277 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271294 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271309 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271326 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271341 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271357 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271369 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271383 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271400 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271412 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271425 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271437 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271448 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.271459 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276712 4575 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276761 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276780 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276797 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276812 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276825 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276840 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276854 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276866 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276877 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276890 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276903 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276917 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276933 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276946 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276957 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276968 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276982 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.276994 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277006 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277017 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277029 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277042 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277053 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277064 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277076 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277088 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277101 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277115 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277127 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277138 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277151 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277163 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277175 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277187 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277200 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277213 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277227 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277243 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277268 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277284 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277300 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277313 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277327 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277341 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277354 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277392 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277405 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277418 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277431 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277444 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277457 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277472 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277485 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277498 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277511 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277524 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277536 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277549 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277561 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277575 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277606 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277620 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277632 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277645 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277659 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277671 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277683 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277696 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277708 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277722 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277735 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277748 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277759 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277772 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277784 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277800 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277812 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277824 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277836 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277850 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277864 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277877 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277892 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277904 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277916 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277932 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277948 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277960 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277973 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277985 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.277998 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278011 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278025 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278038 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278050 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278063 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278074 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278096 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278109 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278123 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278135 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278149 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278161 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278171 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278183 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278195 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278207 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278219 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278230 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278242 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278252 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278263 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278274 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278286 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278299 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278310 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278322 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278337 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278349 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278360 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278371 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278383 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278395 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278408 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278420 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278431 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278444 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278455 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278466 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278478 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278492 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278510 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278521 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278534 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278547 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278558 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278569 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.278579 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.279219 4575 manager.go:324] Recovery completed Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.279223 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.280771 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.280859 4575 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.280879 4575 reconstruct.go:97] "Volume reconstruction finished" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.280894 4575 reconciler.go:26] "Reconciler: start to sync state" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.286685 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.288395 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.288436 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.288450 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.289320 4575 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.289378 4575 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.289404 4575 state_mem.go:36] "Initialized new in-memory state store" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.304726 4575 policy_none.go:49] "None policy: Start" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.306392 4575 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.306416 4575 state_mem.go:35] "Initializing new in-memory state store" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.306530 4575 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.308397 4575 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.308648 4575 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.308693 4575 kubelet.go:2335] "Starting kubelet main sync loop" Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.308936 4575 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.310225 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.310268 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.354558 4575 manager.go:334] "Starting Device Plugin manager" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.354639 4575 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.354686 4575 server.go:79] "Starting device plugin registration server" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.355061 4575 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.355109 4575 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.355269 4575 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.355335 4575 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.355342 4575 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.361559 4575 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.409663 4575 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.409742 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.410382 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.410420 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.410428 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.410566 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.410690 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.410719 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.411343 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.411360 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.411369 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.411394 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.411417 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.411427 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.411524 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.411750 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.411780 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412120 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412149 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412160 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412247 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412430 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412453 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412472 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412478 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412486 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412734 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412746 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412753 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.412818 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413087 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413124 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413262 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413272 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413430 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413452 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413462 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413616 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413639 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413961 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413985 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.413994 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.414181 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.414194 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.414201 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.455398 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.456415 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.456443 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.456456 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.456478 4575 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.456916 4575 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.199:6443: connect: connection refused" node="crc" Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.460345 4575 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.199:6443: connect: connection refused" interval="400ms" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483369 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483420 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483446 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483469 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483489 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483508 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483527 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483549 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483568 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483618 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483648 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483679 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483717 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483742 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.483762 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585097 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585146 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585165 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585184 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585220 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585242 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585267 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585289 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585308 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585327 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585345 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585345 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585423 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585453 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585387 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585496 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585524 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585518 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585571 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585626 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585583 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585679 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585727 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585760 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585758 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585771 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585786 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585552 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585657 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.585827 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.657349 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.658916 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.658957 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.658972 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.659022 4575 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.659560 4575 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.199:6443: connect: connection refused" node="crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.743198 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.765389 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.781179 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.788830 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: I1004 04:34:09.792748 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.850570 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-9cdad6f60ca6558395e78086af8daeec8fc5e6ca77db247de30b027de4c9bc72 WatchSource:0}: Error finding container 9cdad6f60ca6558395e78086af8daeec8fc5e6ca77db247de30b027de4c9bc72: Status 404 returned error can't find the container with id 9cdad6f60ca6558395e78086af8daeec8fc5e6ca77db247de30b027de4c9bc72 Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.851240 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ef0227203dd2dee8ac918bd513efa6455118dc12948dea35c6f28f392636eab1 WatchSource:0}: Error finding container ef0227203dd2dee8ac918bd513efa6455118dc12948dea35c6f28f392636eab1: Status 404 returned error can't find the container with id ef0227203dd2dee8ac918bd513efa6455118dc12948dea35c6f28f392636eab1 Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.852296 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b63041a9d21fba1e7135b1a28eaf4f62f5f85f3a0b5e967e99d82f58f4e5fd4e WatchSource:0}: Error finding container b63041a9d21fba1e7135b1a28eaf4f62f5f85f3a0b5e967e99d82f58f4e5fd4e: Status 404 returned error can't find the container with id b63041a9d21fba1e7135b1a28eaf4f62f5f85f3a0b5e967e99d82f58f4e5fd4e Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.854552 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-57d706d7cdc1e9754fa1956b6b1eccb6c158487d3f950f14b586354e881f4530 WatchSource:0}: Error finding container 57d706d7cdc1e9754fa1956b6b1eccb6c158487d3f950f14b586354e881f4530: Status 404 returned error can't find the container with id 57d706d7cdc1e9754fa1956b6b1eccb6c158487d3f950f14b586354e881f4530 Oct 04 04:34:09 crc kubenswrapper[4575]: W1004 04:34:09.855989 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a13d67cbb7bd03f33a715cd776448d8bca74604900e58fd6046242816dbf40eb WatchSource:0}: Error finding container a13d67cbb7bd03f33a715cd776448d8bca74604900e58fd6046242816dbf40eb: Status 404 returned error can't find the container with id a13d67cbb7bd03f33a715cd776448d8bca74604900e58fd6046242816dbf40eb Oct 04 04:34:09 crc kubenswrapper[4575]: E1004 04:34:09.861085 4575 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.199:6443: connect: connection refused" interval="800ms" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.060018 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.061605 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.061657 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.061669 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.061694 4575 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:34:10 crc kubenswrapper[4575]: E1004 04:34:10.062141 4575 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.199:6443: connect: connection refused" node="crc" Oct 04 04:34:10 crc kubenswrapper[4575]: W1004 04:34:10.122782 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:10 crc kubenswrapper[4575]: E1004 04:34:10.122886 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:10 crc kubenswrapper[4575]: W1004 04:34:10.144017 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:10 crc kubenswrapper[4575]: E1004 04:34:10.144101 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.251314 4575 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.313621 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b63041a9d21fba1e7135b1a28eaf4f62f5f85f3a0b5e967e99d82f58f4e5fd4e"} Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.314810 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ef0227203dd2dee8ac918bd513efa6455118dc12948dea35c6f28f392636eab1"} Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.316489 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9cdad6f60ca6558395e78086af8daeec8fc5e6ca77db247de30b027de4c9bc72"} Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.317504 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a13d67cbb7bd03f33a715cd776448d8bca74604900e58fd6046242816dbf40eb"} Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.318728 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"57d706d7cdc1e9754fa1956b6b1eccb6c158487d3f950f14b586354e881f4530"} Oct 04 04:34:10 crc kubenswrapper[4575]: W1004 04:34:10.572716 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:10 crc kubenswrapper[4575]: E1004 04:34:10.573109 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:10 crc kubenswrapper[4575]: E1004 04:34:10.661925 4575 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.199:6443: connect: connection refused" interval="1.6s" Oct 04 04:34:10 crc kubenswrapper[4575]: W1004 04:34:10.694657 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:10 crc kubenswrapper[4575]: E1004 04:34:10.694767 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.862629 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.863867 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.863907 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.863919 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4575]: I1004 04:34:10.864123 4575 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:34:10 crc kubenswrapper[4575]: E1004 04:34:10.864535 4575 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.199:6443: connect: connection refused" node="crc" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.251895 4575 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.323471 4575 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d" exitCode=0 Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.323547 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d"} Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.323684 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.324768 4575 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="56762753ddca856cbb3ffcd278c916cd3eaceb1a98c370d4dbcb2995a4c14551" exitCode=0 Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.324891 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.324975 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.325019 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.325036 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.325213 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"56762753ddca856cbb3ffcd278c916cd3eaceb1a98c370d4dbcb2995a4c14551"} Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.325539 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.325559 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.325567 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.327103 4575 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="34d8b0568d521140f52aeef7d7dbd8adf289151877bf4866df15adcf1a690922" exitCode=0 Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.327171 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.327172 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"34d8b0568d521140f52aeef7d7dbd8adf289151877bf4866df15adcf1a690922"} Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.327218 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.328036 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.328074 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.328084 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.328977 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.328994 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.329004 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.332007 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01"} Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.332060 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f"} Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.332088 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af"} Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.334090 4575 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772" exitCode=0 Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.334142 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772"} Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.334155 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.334870 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.334896 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4575]: I1004 04:34:11.334905 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4575]: W1004 04:34:11.868985 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:11 crc kubenswrapper[4575]: E1004 04:34:11.869090 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:12 crc kubenswrapper[4575]: W1004 04:34:12.199323 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:12 crc kubenswrapper[4575]: E1004 04:34:12.199427 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.251716 4575 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:12 crc kubenswrapper[4575]: E1004 04:34:12.262409 4575 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.199:6443: connect: connection refused" interval="3.2s" Oct 04 04:34:12 crc kubenswrapper[4575]: W1004 04:34:12.332555 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:12 crc kubenswrapper[4575]: E1004 04:34:12.332645 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.339390 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.339432 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.339443 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.339451 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.341698 4575 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="763c09570e509b049afbc322da36072f418f80432fc64dee763c74b826805822" exitCode=0 Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.341761 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"763c09570e509b049afbc322da36072f418f80432fc64dee763c74b826805822"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.341786 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.342859 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.342881 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.342889 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.344721 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"808a4019218d0e8fe18a085a6027b53da0cccddb5ca08e21452014055b3f4e53"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.344799 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.345371 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.345387 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.345396 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.347265 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.347339 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.348173 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.348195 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.348202 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.351245 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.351270 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.351279 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc"} Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.351348 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.351964 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.351984 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.351992 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.464926 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.466381 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.466417 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.466428 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4575]: I1004 04:34:12.466448 4575 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:34:12 crc kubenswrapper[4575]: E1004 04:34:12.466911 4575 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.199:6443: connect: connection refused" node="crc" Oct 04 04:34:12 crc kubenswrapper[4575]: W1004 04:34:12.624925 4575 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.199:6443: connect: connection refused Oct 04 04:34:12 crc kubenswrapper[4575]: E1004 04:34:12.625022 4575 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.199:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.355922 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17"} Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.356067 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.356893 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.356921 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.356933 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.358768 4575 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="974491db5db16e05d1724d1649e3ddef7dacfda89412ea9fa8114882a93d87c9" exitCode=0 Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.358831 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.359103 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"974491db5db16e05d1724d1649e3ddef7dacfda89412ea9fa8114882a93d87c9"} Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.359150 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.359340 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.359420 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.359538 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.359888 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.359903 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.359910 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.361424 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.361465 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.361484 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.361685 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.361753 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.361777 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.362333 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.362371 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4575]: I1004 04:34:13.362385 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.040123 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.048236 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.364138 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.364741 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.364776 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7110dda2fde5e5727d5164bba3095dd2c4fb72a47ea83fd5685a61e3eab09841"} Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.364803 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"94cfb6f9aa2a7971613ce6eb1d6e3dd6769e59178b1c2a3abc3f9813e74a5eda"} Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.364817 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"94d397efaf62cb3c1510b3ab396d1efe96755c089882de5e5bb528e7e27406ce"} Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.364828 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2324c47409970f6d0d6987416be0be92476c01bce8e9095f77f3db1ab43496e0"} Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.364839 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bdde181cec7a5f92c6ef7871aa9db54a7601367f04cfc335046c93e3087bc4d3"} Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.364917 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365050 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365344 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365366 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365374 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365514 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365555 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365567 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365893 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365925 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.365941 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4575]: I1004 04:34:14.748519 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.081378 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.367276 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.367946 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.368077 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.367374 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.369297 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.369342 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.369357 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.369547 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.369571 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.369607 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.370233 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.370297 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.370322 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.667873 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.669068 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.669129 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.669145 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4575]: I1004 04:34:15.669178 4575 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:34:16 crc kubenswrapper[4575]: I1004 04:34:16.369854 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:16 crc kubenswrapper[4575]: I1004 04:34:16.371799 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4575]: I1004 04:34:16.371824 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4575]: I1004 04:34:16.371835 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.330293 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.331489 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.331698 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.333634 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.333679 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.333688 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.444522 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.444732 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.445701 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.445741 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.445751 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.921233 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.921451 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.922550 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.922580 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4575]: I1004 04:34:17.922621 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4575]: I1004 04:34:18.124952 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:18 crc kubenswrapper[4575]: I1004 04:34:18.125121 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:18 crc kubenswrapper[4575]: I1004 04:34:18.126151 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4575]: I1004 04:34:18.126201 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4575]: I1004 04:34:18.126209 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4575]: I1004 04:34:19.300551 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 04 04:34:19 crc kubenswrapper[4575]: I1004 04:34:19.300845 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:19 crc kubenswrapper[4575]: I1004 04:34:19.302395 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4575]: I1004 04:34:19.302434 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4575]: I1004 04:34:19.302446 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4575]: E1004 04:34:19.361831 4575 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.318419 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.318570 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.319557 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.319608 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.319619 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.323437 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.330468 4575 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.330542 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.381299 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.382913 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.383052 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4575]: I1004 04:34:20.383100 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.253811 4575 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.390128 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.392075 4575 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17" exitCode=255 Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.392118 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17"} Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.392293 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.393130 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.393165 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.393178 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.393647 4575 scope.go:117] "RemoveContainer" containerID="cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17" Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.899449 4575 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.899525 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.908634 4575 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 04 04:34:23 crc kubenswrapper[4575]: I1004 04:34:23.908891 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 04:34:24 crc kubenswrapper[4575]: I1004 04:34:24.397742 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 04:34:24 crc kubenswrapper[4575]: I1004 04:34:24.400395 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098"} Oct 04 04:34:24 crc kubenswrapper[4575]: I1004 04:34:24.400642 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:24 crc kubenswrapper[4575]: I1004 04:34:24.402206 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4575]: I1004 04:34:24.402248 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4575]: I1004 04:34:24.402260 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4575]: I1004 04:34:24.757054 4575 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]log ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]etcd ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/priority-and-fairness-filter ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-apiextensions-informers ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-apiextensions-controllers ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/crd-informer-synced ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-system-namespaces-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 04 04:34:24 crc kubenswrapper[4575]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 04 04:34:24 crc kubenswrapper[4575]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/bootstrap-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/start-kube-aggregator-informers ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/apiservice-registration-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/apiservice-discovery-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]autoregister-completion ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/apiservice-openapi-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 04 04:34:24 crc kubenswrapper[4575]: livez check failed Oct 04 04:34:24 crc kubenswrapper[4575]: I1004 04:34:24.757159 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:34:27 crc kubenswrapper[4575]: I1004 04:34:27.950425 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 04 04:34:27 crc kubenswrapper[4575]: I1004 04:34:27.951345 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:27 crc kubenswrapper[4575]: I1004 04:34:27.952635 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:27 crc kubenswrapper[4575]: I1004 04:34:27.952786 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:27 crc kubenswrapper[4575]: I1004 04:34:27.952911 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:27 crc kubenswrapper[4575]: I1004 04:34:27.963760 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.409697 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.410446 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.410477 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.410487 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:28 crc kubenswrapper[4575]: E1004 04:34:28.870796 4575 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.872244 4575 trace.go:236] Trace[1047969388]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 04:34:17.507) (total time: 11364ms): Oct 04 04:34:28 crc kubenswrapper[4575]: Trace[1047969388]: ---"Objects listed" error: 11364ms (04:34:28.872) Oct 04 04:34:28 crc kubenswrapper[4575]: Trace[1047969388]: [11.364262114s] [11.364262114s] END Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.872701 4575 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.876140 4575 trace.go:236] Trace[1835910337]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 04:34:18.495) (total time: 10380ms): Oct 04 04:34:28 crc kubenswrapper[4575]: Trace[1835910337]: ---"Objects listed" error: 10380ms (04:34:28.876) Oct 04 04:34:28 crc kubenswrapper[4575]: Trace[1835910337]: [10.380872691s] [10.380872691s] END Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.876167 4575 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.877007 4575 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.877032 4575 trace.go:236] Trace[949752141]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 04:34:15.694) (total time: 13181ms): Oct 04 04:34:28 crc kubenswrapper[4575]: Trace[949752141]: ---"Objects listed" error: 13181ms (04:34:28.876) Oct 04 04:34:28 crc kubenswrapper[4575]: Trace[949752141]: [13.181992141s] [13.181992141s] END Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.877553 4575 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.877673 4575 trace.go:236] Trace[845826444]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 04:34:17.414) (total time: 11463ms): Oct 04 04:34:28 crc kubenswrapper[4575]: Trace[845826444]: ---"Objects listed" error: 11463ms (04:34:28.877) Oct 04 04:34:28 crc kubenswrapper[4575]: Trace[845826444]: [11.463158571s] [11.463158571s] END Oct 04 04:34:28 crc kubenswrapper[4575]: I1004 04:34:28.877756 4575 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 04 04:34:28 crc kubenswrapper[4575]: E1004 04:34:28.877911 4575 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.244091 4575 apiserver.go:52] "Watching apiserver" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.247169 4575 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.247437 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.247769 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.247866 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.247872 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.248039 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.248137 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.248153 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.248554 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.248666 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.248742 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.249657 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.250219 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.250652 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.250704 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.250828 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.251064 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.253041 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.253022 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.253080 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.253738 4575 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278627 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278677 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278697 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278714 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278730 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278744 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278759 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278773 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278787 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278801 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278815 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278832 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278849 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278864 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278878 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278892 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278907 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278923 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278937 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278950 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278964 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278980 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.278993 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279008 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279025 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279039 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279055 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279150 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279167 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279183 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279198 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279215 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279229 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279244 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279314 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279340 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279354 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279370 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279384 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279399 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279416 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279432 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279447 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279464 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279481 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279496 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279511 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279525 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279543 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279558 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279576 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279579 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279625 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279641 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279656 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279671 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279687 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279707 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279742 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279757 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279771 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279788 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279804 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279818 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279834 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279853 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279870 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279884 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279900 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279914 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279932 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279948 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279964 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279979 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279995 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280009 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280024 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280040 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280056 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280072 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280088 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280105 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280120 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280135 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280167 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280191 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280207 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280221 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280239 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280253 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280267 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280282 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280296 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280312 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280328 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280343 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280359 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280377 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280392 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280407 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280423 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280438 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280520 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280537 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280553 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280573 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280608 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280624 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280639 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280654 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280669 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280684 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280698 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280715 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280730 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280745 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280760 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280776 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280793 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280809 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280825 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280842 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280857 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280873 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280889 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280906 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280921 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280937 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280953 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280971 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280988 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281004 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281019 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281037 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281054 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281072 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281088 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281103 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281122 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281138 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281154 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281169 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281185 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281200 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281215 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281232 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281249 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281265 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281281 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281297 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281313 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281328 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281345 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281361 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281377 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281392 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281408 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281424 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281440 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281455 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281471 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281486 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281502 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281518 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281533 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281549 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281569 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281632 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281673 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281694 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281712 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281729 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281746 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281763 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281780 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281795 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281813 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281833 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281849 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281866 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281881 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281899 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281914 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281929 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281945 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281962 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281980 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281997 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282015 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282033 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282049 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282065 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282081 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282099 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282118 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282135 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282169 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282189 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282212 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282232 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282251 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282271 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282286 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282303 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282321 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282353 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282374 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282392 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282411 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282429 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282467 4575 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279770 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297790 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.279887 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280098 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280384 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280569 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280754 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280794 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.280901 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281005 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281047 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281178 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281177 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281297 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281645 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281658 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281801 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281937 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.281989 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282065 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282168 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282185 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282303 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282354 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.282515 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.283731 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.283971 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.284046 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.284418 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.284425 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.284488 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.284685 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.284716 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.284855 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.285037 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.285116 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.285122 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.285252 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.286764 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.286798 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.286957 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.287099 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.287371 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.287454 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:34:29.787436138 +0000 UTC m=+21.115994952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.287478 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.287610 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.288228 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.288245 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.288412 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.288489 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.288620 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.288775 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.288874 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.288989 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.289139 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.289669 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.292148 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.293919 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.295285 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.296040 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.296578 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.296970 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297216 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297395 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297463 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297579 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297642 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297752 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297812 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297859 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.297951 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.298084 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.298216 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.298628 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.298713 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.299074 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.299074 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.299178 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.299537 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.299571 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.299690 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.299792 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.299960 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.300155 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.300368 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.300571 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.300684 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.330725 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.337137 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.337370 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.337561 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.337779 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.337929 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.338006 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.338100 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.338321 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.338636 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.338989 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.340086 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.341157 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.345968 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.346201 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.346457 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.346617 4575 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.347075 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.347132 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.347364 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.347627 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.348068 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.348361 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.348597 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.352532 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.352769 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.352810 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.353369 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.353955 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.354218 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.354242 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.354669 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.355150 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.355223 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.355573 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.356156 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.356632 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.356815 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.356856 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.357065 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.357179 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.357365 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.357667 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.357952 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.358320 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.358414 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.358460 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.358617 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.358813 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.361905 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.362195 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.362488 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.362604 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.362818 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.363026 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.363040 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.363077 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.363402 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.363663 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.363699 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.363720 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.363776 4575 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.363932 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.363970 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:29.863953875 +0000 UTC m=+21.192512689 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.364404 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.364473 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.364824 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.365225 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.365508 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.366491 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.367092 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.368672 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.368897 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.369459 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.370289 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.371192 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.371323 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.371515 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.371524 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.372155 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.372770 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.372984 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.373550 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.373634 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.374214 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.374431 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.374616 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.374711 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.374924 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.374967 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.375026 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.375490 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.379568 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.380374 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.381613 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.381787 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.381807 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.381902 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.382100 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.382243 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.382257 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.382294 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.382341 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.382431 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.382461 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.382749 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.382887 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.384028 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.384122 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.384275 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.384344 4575 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.384459 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:29.88443388 +0000 UTC m=+21.212992754 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.384576 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.384702 4575 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.384798 4575 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.384871 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.384940 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.387979 4575 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.388088 4575 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.388168 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.388226 4575 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.388283 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389254 4575 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389278 4575 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389289 4575 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389300 4575 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389311 4575 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389321 4575 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389331 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389343 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389354 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389401 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389414 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389426 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389439 4575 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389451 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389463 4575 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389473 4575 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389484 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389495 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389507 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389518 4575 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389530 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389542 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389552 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389564 4575 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389575 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389609 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389620 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389632 4575 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389643 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389656 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389667 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389826 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389841 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389852 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389863 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389875 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389887 4575 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389898 4575 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389909 4575 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389944 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389958 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389970 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389982 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389994 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390005 4575 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390015 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390025 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390008 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390037 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390139 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390155 4575 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390164 4575 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390173 4575 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390196 4575 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390208 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390219 4575 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390229 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390239 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390250 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390259 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390268 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390276 4575 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390284 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390293 4575 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390301 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390309 4575 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390317 4575 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390347 4575 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390358 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390366 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390374 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390383 4575 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390392 4575 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.386110 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.388822 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.388835 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389049 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389229 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.386678 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390401 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390559 4575 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390573 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390609 4575 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390623 4575 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390635 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390647 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390659 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390691 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390704 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390715 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389594 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390726 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.389898 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390737 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390772 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390785 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390798 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390809 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390820 4575 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390855 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390867 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390878 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390889 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390900 4575 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390932 4575 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390944 4575 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390956 4575 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390966 4575 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390977 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391011 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391024 4575 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391037 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391049 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391061 4575 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391092 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391105 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391117 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391128 4575 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391139 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.390342 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391171 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391183 4575 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391195 4575 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391206 4575 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391254 4575 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391266 4575 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391278 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391289 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391299 4575 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391331 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391345 4575 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391360 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391371 4575 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391382 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391415 4575 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391426 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391438 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391450 4575 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391461 4575 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391493 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391505 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391516 4575 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391528 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391539 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391550 4575 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391633 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391651 4575 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391663 4575 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391675 4575 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391688 4575 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391723 4575 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391736 4575 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.391747 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.392516 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.393156 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.394894 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.395620 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.395931 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.396572 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.396975 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.397634 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.397687 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.399558 4575 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.399749 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.402233 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.402520 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.403524 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.404723 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.406828 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.408130 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.409224 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.411984 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.413882 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.415776 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.416413 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.416662 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.419028 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.420604 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.422797 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.440695 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.440901 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.440965 4575 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.441094 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:29.941074088 +0000 UTC m=+21.269632903 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.441353 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.441430 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.441485 4575 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.441599 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:29.941575963 +0000 UTC m=+21.270134777 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.442236 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.442678 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.443428 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.444951 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.445474 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.448823 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.449867 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.451719 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.453650 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.455272 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.456250 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.457920 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-2sdsl"] Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.458399 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2sdsl" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.460105 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.463495 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.473219 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.479044 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.479407 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.480604 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.481390 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492139 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e1f25b10-0668-41ff-9197-c1fed6995ca6-hosts-file\") pod \"node-resolver-2sdsl\" (UID: \"e1f25b10-0668-41ff-9197-c1fed6995ca6\") " pod="openshift-dns/node-resolver-2sdsl" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492195 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjc4f\" (UniqueName: \"kubernetes.io/projected/e1f25b10-0668-41ff-9197-c1fed6995ca6-kube-api-access-tjc4f\") pod \"node-resolver-2sdsl\" (UID: \"e1f25b10-0668-41ff-9197-c1fed6995ca6\") " pod="openshift-dns/node-resolver-2sdsl" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492233 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492266 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492276 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492286 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492294 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492302 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492310 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492318 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492326 4575 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492334 4575 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492342 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492349 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492357 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492366 4575 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492374 4575 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492383 4575 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492391 4575 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492399 4575 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492407 4575 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492416 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492424 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492432 4575 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492443 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492452 4575 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492460 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492468 4575 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492476 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492483 4575 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492491 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492499 4575 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492507 4575 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492514 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492522 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492530 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492538 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492547 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492556 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492564 4575 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492572 4575 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492579 4575 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492601 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.492634 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.497951 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.510535 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.518273 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.525928 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.528967 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.529452 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.542873 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.555535 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.555705 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.561125 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.569798 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.570078 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: W1004 04:34:29.576731 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-ca7460792c85b0b0988b66ec8eca690fd8f2395c9b377d9d1e8f995b80976c68 WatchSource:0}: Error finding container ca7460792c85b0b0988b66ec8eca690fd8f2395c9b377d9d1e8f995b80976c68: Status 404 returned error can't find the container with id ca7460792c85b0b0988b66ec8eca690fd8f2395c9b377d9d1e8f995b80976c68 Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.580462 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.588782 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.592842 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e1f25b10-0668-41ff-9197-c1fed6995ca6-hosts-file\") pod \"node-resolver-2sdsl\" (UID: \"e1f25b10-0668-41ff-9197-c1fed6995ca6\") " pod="openshift-dns/node-resolver-2sdsl" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.592886 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjc4f\" (UniqueName: \"kubernetes.io/projected/e1f25b10-0668-41ff-9197-c1fed6995ca6-kube-api-access-tjc4f\") pod \"node-resolver-2sdsl\" (UID: \"e1f25b10-0668-41ff-9197-c1fed6995ca6\") " pod="openshift-dns/node-resolver-2sdsl" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.593137 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e1f25b10-0668-41ff-9197-c1fed6995ca6-hosts-file\") pod \"node-resolver-2sdsl\" (UID: \"e1f25b10-0668-41ff-9197-c1fed6995ca6\") " pod="openshift-dns/node-resolver-2sdsl" Oct 04 04:34:29 crc kubenswrapper[4575]: W1004 04:34:29.593825 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-6b1dc8b7ef3fb63c912c1e5a7186ff8cdab1601caa9aca7ac85cf0c80bb09509 WatchSource:0}: Error finding container 6b1dc8b7ef3fb63c912c1e5a7186ff8cdab1601caa9aca7ac85cf0c80bb09509: Status 404 returned error can't find the container with id 6b1dc8b7ef3fb63c912c1e5a7186ff8cdab1601caa9aca7ac85cf0c80bb09509 Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.599056 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.614162 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjc4f\" (UniqueName: \"kubernetes.io/projected/e1f25b10-0668-41ff-9197-c1fed6995ca6-kube-api-access-tjc4f\") pod \"node-resolver-2sdsl\" (UID: \"e1f25b10-0668-41ff-9197-c1fed6995ca6\") " pod="openshift-dns/node-resolver-2sdsl" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.614447 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.625053 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.641642 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.752651 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.753265 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.760281 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.771516 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.782814 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.789372 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2sdsl" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.790787 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-pl4jq"] Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.791160 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.796322 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.796465 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:34:30.796439418 +0000 UTC m=+22.124998232 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.797769 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.798024 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 04 04:34:29 crc kubenswrapper[4575]: W1004 04:34:29.799515 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1f25b10_0668_41ff_9197_c1fed6995ca6.slice/crio-9ec58d98db56b44ed307be08f8ce2cd055433533ca6c15d00eb0eb8c5a1c9dc7 WatchSource:0}: Error finding container 9ec58d98db56b44ed307be08f8ce2cd055433533ca6c15d00eb0eb8c5a1c9dc7: Status 404 returned error can't find the container with id 9ec58d98db56b44ed307be08f8ce2cd055433533ca6c15d00eb0eb8c5a1c9dc7 Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.800310 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.800552 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.800667 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.804121 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.831501 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.853048 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.886531 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.896684 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.896721 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/df9f335e-3c50-4cbf-ad47-491eb2486de8-rootfs\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.896740 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.896762 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df9f335e-3c50-4cbf-ad47-491eb2486de8-mcd-auth-proxy-config\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.896777 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr2vw\" (UniqueName: \"kubernetes.io/projected/df9f335e-3c50-4cbf-ad47-491eb2486de8-kube-api-access-pr2vw\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.896814 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df9f335e-3c50-4cbf-ad47-491eb2486de8-proxy-tls\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.896826 4575 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.896897 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:30.896880431 +0000 UTC m=+22.225439245 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.896938 4575 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.896980 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:30.896967403 +0000 UTC m=+22.225526207 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.916716 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.929647 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.941478 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.947376 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.961433 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.973500 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.980325 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.988170 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.997872 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.997921 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df9f335e-3c50-4cbf-ad47-491eb2486de8-proxy-tls\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.997961 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/df9f335e-3c50-4cbf-ad47-491eb2486de8-rootfs\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.997982 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df9f335e-3c50-4cbf-ad47-491eb2486de8-mcd-auth-proxy-config\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.998001 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr2vw\" (UniqueName: \"kubernetes.io/projected/df9f335e-3c50-4cbf-ad47-491eb2486de8-kube-api-access-pr2vw\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.998025 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.998050 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/df9f335e-3c50-4cbf-ad47-491eb2486de8-rootfs\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.998056 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.998121 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.998136 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.998138 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.998148 4575 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.998151 4575 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.998191 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:30.998175808 +0000 UTC m=+22.326734622 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:29 crc kubenswrapper[4575]: E1004 04:34:29.998209 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:30.998200969 +0000 UTC m=+22.326759783 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:29 crc kubenswrapper[4575]: I1004 04:34:29.998973 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df9f335e-3c50-4cbf-ad47-491eb2486de8-mcd-auth-proxy-config\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.000848 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.023243 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.037000 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.054861 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.077805 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df9f335e-3c50-4cbf-ad47-491eb2486de8-proxy-tls\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.078096 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr2vw\" (UniqueName: \"kubernetes.io/projected/df9f335e-3c50-4cbf-ad47-491eb2486de8-kube-api-access-pr2vw\") pod \"machine-config-daemon-pl4jq\" (UID: \"df9f335e-3c50-4cbf-ad47-491eb2486de8\") " pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.113261 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.124203 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf9f335e_3c50_4cbf_ad47_491eb2486de8.slice/crio-4221da65a3f57f6971a0ff8448eec112f52150d418bbcc1c908d7179bf441846 WatchSource:0}: Error finding container 4221da65a3f57f6971a0ff8448eec112f52150d418bbcc1c908d7179bf441846: Status 404 returned error can't find the container with id 4221da65a3f57f6971a0ff8448eec112f52150d418bbcc1c908d7179bf441846 Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.237630 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-kmr4x"] Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.238252 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.239724 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hqg47"] Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.240653 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-sj5hq"] Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.240672 4575 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.240733 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.240824 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.241095 4575 reflector.go:561] object-"openshift-multus"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.241120 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.241477 4575 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.241510 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.241476 4575 reflector.go:561] object-"openshift-multus"/"default-cni-sysctl-allowlist": failed to list *v1.ConfigMap: configmaps "default-cni-sysctl-allowlist" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.241540 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"default-cni-sysctl-allowlist\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.241632 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.241982 4575 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.242016 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.242540 4575 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovnkube-config": failed to list *v1.ConfigMap: configmaps "ovnkube-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.242640 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnkube-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.242645 4575 reflector.go:561] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.242605 4575 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-pwtwl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.242752 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-pwtwl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-node-dockercfg-pwtwl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.242760 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.244855 4575 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": failed to list *v1.Secret: secrets "ovn-node-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.244886 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-node-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.244891 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 04 04:34:30 crc kubenswrapper[4575]: W1004 04:34:30.244930 4575 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": failed to list *v1.ConfigMap: configmaps "ovnkube-script-lib" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.244956 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnkube-script-lib\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.245202 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.245372 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.245511 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.258243 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.276456 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.293876 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.300721 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-system-cni-dir\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.300793 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-cni-binary-copy\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.300833 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-run-netns\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.300876 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-run-multus-certs\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.300897 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-etc-kubernetes\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.300952 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-slash\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.300976 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-log-socket\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.300997 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/984bf56e-d83a-4206-a887-116d728163d1-ovn-node-metrics-cert\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301038 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-systemd-units\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301059 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-var-lib-cni-multus\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301081 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-cnibin\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301124 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-kubelet\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301143 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-node-log\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301163 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-netns\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301205 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-netd\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301226 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301247 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-binary-copy\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301289 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-var-lib-openvswitch\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301317 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-ovn\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301360 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-run-k8s-cni-cncf-io\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301381 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-var-lib-kubelet\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301400 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-daemon-config\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301443 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8b6z\" (UniqueName: \"kubernetes.io/projected/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-kube-api-access-w8b6z\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301471 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-system-cni-dir\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301511 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-systemd\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301532 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-etc-openvswitch\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301610 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-bin\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301639 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-os-release\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301685 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-config\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301707 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-script-lib\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301727 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfjxd\" (UniqueName: \"kubernetes.io/projected/984bf56e-d83a-4206-a887-116d728163d1-kube-api-access-sfjxd\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301768 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-cnibin\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301787 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-conf-dir\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301807 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-ovn-kubernetes\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301849 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-hostroot\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301882 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301902 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-openvswitch\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301944 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-cni-dir\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.301965 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-var-lib-cni-bin\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.302007 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.302053 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-os-release\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.302096 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-socket-dir-parent\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.302119 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpn2j\" (UniqueName: \"kubernetes.io/projected/02c08151-21fd-4ed4-abd7-3ec7e427df35-kube-api-access-zpn2j\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.302172 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-env-overrides\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.309818 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.310025 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.352658 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.402834 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-systemd-units\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.402879 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-var-lib-cni-multus\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.402899 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-cnibin\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.402916 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-kubelet\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.402934 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-node-log\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.402953 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.402973 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-netns\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.402988 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-netd\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403004 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-binary-copy\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403020 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-var-lib-openvswitch\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403038 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-ovn\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403047 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403085 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-var-lib-kubelet\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403122 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-var-lib-openvswitch\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403123 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-kubelet\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403144 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-ovn\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403008 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-var-lib-cni-multus\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403152 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-netns\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403054 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-var-lib-kubelet\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403218 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-systemd-units\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403242 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-netd\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403242 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-node-log\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403309 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-cnibin\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403314 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-daemon-config\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403447 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-run-k8s-cni-cncf-io\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403473 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8b6z\" (UniqueName: \"kubernetes.io/projected/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-kube-api-access-w8b6z\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403489 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-system-cni-dir\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403504 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-systemd\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403518 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-etc-openvswitch\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403532 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-bin\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403543 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-system-cni-dir\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403579 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-systemd\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403547 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-script-lib\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403632 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-etc-openvswitch\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403645 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfjxd\" (UniqueName: \"kubernetes.io/projected/984bf56e-d83a-4206-a887-116d728163d1-kube-api-access-sfjxd\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403664 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-bin\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403682 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-os-release\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403698 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-run-k8s-cni-cncf-io\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403700 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-config\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403729 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-cnibin\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403743 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-conf-dir\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403758 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-ovn-kubernetes\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403773 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-hostroot\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403788 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403793 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-cnibin\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403803 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-openvswitch\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403818 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-cni-dir\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403824 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-hostroot\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403835 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-var-lib-cni-bin\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403833 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-ovn-kubernetes\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403862 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403870 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-openvswitch\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403855 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-var-lib-cni-bin\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403895 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpn2j\" (UniqueName: \"kubernetes.io/projected/02c08151-21fd-4ed4-abd7-3ec7e427df35-kube-api-access-zpn2j\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403926 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-conf-dir\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403931 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-cni-dir\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.403988 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-os-release\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404081 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-os-release\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404101 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-socket-dir-parent\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404134 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-daemon-config\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404181 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-multus-socket-dir-parent\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404129 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-env-overrides\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404184 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-os-release\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404231 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-system-cni-dir\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404246 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-cni-binary-copy\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404272 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-run-netns\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404282 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-system-cni-dir\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404308 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-run-netns\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404327 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-run-multus-certs\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404342 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-host-run-multus-certs\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404388 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-etc-kubernetes\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404411 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-etc-kubernetes\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404424 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-slash\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404443 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-log-socket\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404458 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/984bf56e-d83a-4206-a887-116d728163d1-ovn-node-metrics-cert\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404461 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-slash\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404497 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-log-socket\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404544 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02c08151-21fd-4ed4-abd7-3ec7e427df35-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.404707 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-env-overrides\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.410033 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.423157 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508"} Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.423200 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ca7460792c85b0b0988b66ec8eca690fd8f2395c9b377d9d1e8f995b80976c68"} Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.424563 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"4221da65a3f57f6971a0ff8448eec112f52150d418bbcc1c908d7179bf441846"} Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.425566 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2sdsl" event={"ID":"e1f25b10-0668-41ff-9197-c1fed6995ca6","Type":"ContainerStarted","Data":"a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943"} Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.425624 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2sdsl" event={"ID":"e1f25b10-0668-41ff-9197-c1fed6995ca6","Type":"ContainerStarted","Data":"9ec58d98db56b44ed307be08f8ce2cd055433533ca6c15d00eb0eb8c5a1c9dc7"} Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.426469 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6b1dc8b7ef3fb63c912c1e5a7186ff8cdab1601caa9aca7ac85cf0c80bb09509"} Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.427922 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74"} Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.427948 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2ec9ee0063139f9d21f2cc735cb221339eb1a6ccbf5712e51a3858c305d34cdc"} Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.449068 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.464930 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.477661 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.491798 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.505974 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.515714 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.531930 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.540880 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.550704 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.559375 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.569572 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.579440 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.589185 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.595887 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.608091 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.620406 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.633437 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.644793 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.663137 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.808063 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.808221 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:34:32.808190825 +0000 UTC m=+24.136749699 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.909315 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:30 crc kubenswrapper[4575]: I1004 04:34:30.909432 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.909495 4575 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.909519 4575 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.909603 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:32.909565595 +0000 UTC m=+24.238124469 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:30 crc kubenswrapper[4575]: E1004 04:34:30.909627 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:32.909619577 +0000 UTC m=+24.238178471 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.010818 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.010986 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.011037 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.011077 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.011090 4575 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.011128 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.011149 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.011162 4575 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.011150 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:33.01113352 +0000 UTC m=+24.339692334 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.011228 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:33.011209422 +0000 UTC m=+24.339768286 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.097367 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.232889 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.308998 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.309169 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.309314 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.309498 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.310358 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.313254 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.313965 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.314714 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.315122 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-config\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.315352 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.316041 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.316687 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.317248 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.317801 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.318347 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.318976 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.319559 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.320085 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.320525 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.322888 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.323476 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.324713 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.326541 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.335392 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfjxd\" (UniqueName: \"kubernetes.io/projected/984bf56e-d83a-4206-a887-116d728163d1-kube-api-access-sfjxd\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.403526 4575 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.403645 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-binary-copy podName:02c08151-21fd-4ed4-abd7-3ec7e427df35 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:31.90362267 +0000 UTC m=+23.232181494 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-binary-copy") pod "multus-additional-cni-plugins-kmr4x" (UID: "02c08151-21fd-4ed4-abd7-3ec7e427df35") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.403826 4575 configmap.go:193] Couldn't get configMap openshift-ovn-kubernetes/ovnkube-script-lib: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.403967 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-script-lib podName:984bf56e-d83a-4206-a887-116d728163d1 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:31.903939509 +0000 UTC m=+23.232498323 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovnkube-script-lib" (UniqueName: "kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-script-lib") pod "ovnkube-node-hqg47" (UID: "984bf56e-d83a-4206-a887-116d728163d1") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.403997 4575 configmap.go:193] Couldn't get configMap openshift-multus/default-cni-sysctl-allowlist: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.404033 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-sysctl-allowlist podName:02c08151-21fd-4ed4-abd7-3ec7e427df35 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:31.904026702 +0000 UTC m=+23.232585516 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-sysctl-allowlist" (UniqueName: "kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-sysctl-allowlist") pod "multus-additional-cni-plugins-kmr4x" (UID: "02c08151-21fd-4ed4-abd7-3ec7e427df35") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.404525 4575 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.404636 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-cni-binary-copy podName:31197ccb-a3f9-4130-bca4-ca1c3d3ee065 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:31.904617069 +0000 UTC m=+23.233175883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-cni-binary-copy") pod "multus-sj5hq" (UID: "31197ccb-a3f9-4130-bca4-ca1c3d3ee065") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.404711 4575 secret.go:188] Couldn't get secret openshift-ovn-kubernetes/ovn-node-metrics-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.404769 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984bf56e-d83a-4206-a887-116d728163d1-ovn-node-metrics-cert podName:984bf56e-d83a-4206-a887-116d728163d1 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:31.904752893 +0000 UTC m=+23.233311707 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovn-node-metrics-cert" (UniqueName: "kubernetes.io/secret/984bf56e-d83a-4206-a887-116d728163d1-ovn-node-metrics-cert") pod "ovnkube-node-hqg47" (UID: "984bf56e-d83a-4206-a887-116d728163d1") : failed to sync secret cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.431938 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006"} Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.431996 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db"} Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.434291 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a"} Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.448485 4575 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.448553 4575 projected.go:194] Error preparing data for projected volume kube-api-access-w8b6z for pod openshift-multus/multus-sj5hq: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.448654 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-kube-api-access-w8b6z podName:31197ccb-a3f9-4130-bca4-ca1c3d3ee065 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:31.9486313 +0000 UTC m=+23.277190114 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-w8b6z" (UniqueName: "kubernetes.io/projected/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-kube-api-access-w8b6z") pod "multus-sj5hq" (UID: "31197ccb-a3f9-4130-bca4-ca1c3d3ee065") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.448986 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.451030 4575 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.451061 4575 projected.go:194] Error preparing data for projected volume kube-api-access-zpn2j for pod openshift-multus/multus-additional-cni-plugins-kmr4x: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: E1004 04:34:31.451117 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/02c08151-21fd-4ed4-abd7-3ec7e427df35-kube-api-access-zpn2j podName:02c08151-21fd-4ed4-abd7-3ec7e427df35 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:31.951102192 +0000 UTC m=+23.279661006 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-zpn2j" (UniqueName: "kubernetes.io/projected/02c08151-21fd-4ed4-abd7-3ec7e427df35-kube-api-access-zpn2j") pod "multus-additional-cni-plugins-kmr4x" (UID: "02c08151-21fd-4ed4-abd7-3ec7e427df35") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.463926 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.476000 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.479410 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.492610 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.501475 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.503813 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.515845 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.527173 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.539129 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.550603 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.569167 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.585985 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.597311 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.612659 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.626203 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.638846 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.654047 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.674557 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.687141 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.699657 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.711343 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.723347 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.723865 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.740340 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.752184 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.767463 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.777504 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.795656 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.806610 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.809133 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.840075 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.923891 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-cni-binary-copy\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.923939 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/984bf56e-d83a-4206-a887-116d728163d1-ovn-node-metrics-cert\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.923961 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-binary-copy\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.924014 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-script-lib\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.924223 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.924766 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-binary-copy\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.924949 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02c08151-21fd-4ed4-abd7-3ec7e427df35-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.924987 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-cni-binary-copy\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.925524 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-script-lib\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:31 crc kubenswrapper[4575]: I1004 04:34:31.929334 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/984bf56e-d83a-4206-a887-116d728163d1-ovn-node-metrics-cert\") pod \"ovnkube-node-hqg47\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.025819 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpn2j\" (UniqueName: \"kubernetes.io/projected/02c08151-21fd-4ed4-abd7-3ec7e427df35-kube-api-access-zpn2j\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.025896 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8b6z\" (UniqueName: \"kubernetes.io/projected/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-kube-api-access-w8b6z\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.029410 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpn2j\" (UniqueName: \"kubernetes.io/projected/02c08151-21fd-4ed4-abd7-3ec7e427df35-kube-api-access-zpn2j\") pod \"multus-additional-cni-plugins-kmr4x\" (UID: \"02c08151-21fd-4ed4-abd7-3ec7e427df35\") " pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.029465 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8b6z\" (UniqueName: \"kubernetes.io/projected/31197ccb-a3f9-4130-bca4-ca1c3d3ee065-kube-api-access-w8b6z\") pod \"multus-sj5hq\" (UID: \"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\") " pod="openshift-multus/multus-sj5hq" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.054264 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.066655 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.073454 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sj5hq" Oct 04 04:34:32 crc kubenswrapper[4575]: W1004 04:34:32.106881 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31197ccb_a3f9_4130_bca4_ca1c3d3ee065.slice/crio-71174d8f2736d574c9d8f6ebe15ec0dd695306691957dd27fca8961c00a180b0 WatchSource:0}: Error finding container 71174d8f2736d574c9d8f6ebe15ec0dd695306691957dd27fca8961c00a180b0: Status 404 returned error can't find the container with id 71174d8f2736d574c9d8f6ebe15ec0dd695306691957dd27fca8961c00a180b0 Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.137908 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2q7bk"] Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.138310 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.143623 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.143913 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.144200 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.144366 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.153303 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.171167 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.186528 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.198554 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.208330 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.222380 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.228452 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9cf7e8f8-8031-463c-8267-255938aad822-host\") pod \"node-ca-2q7bk\" (UID: \"9cf7e8f8-8031-463c-8267-255938aad822\") " pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.228485 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9cf7e8f8-8031-463c-8267-255938aad822-serviceca\") pod \"node-ca-2q7bk\" (UID: \"9cf7e8f8-8031-463c-8267-255938aad822\") " pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.228505 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq7tx\" (UniqueName: \"kubernetes.io/projected/9cf7e8f8-8031-463c-8267-255938aad822-kube-api-access-cq7tx\") pod \"node-ca-2q7bk\" (UID: \"9cf7e8f8-8031-463c-8267-255938aad822\") " pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.233868 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.250958 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.264016 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.275163 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.288440 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.302475 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.309280 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:32 crc kubenswrapper[4575]: E1004 04:34:32.309383 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.313972 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.324129 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.329476 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9cf7e8f8-8031-463c-8267-255938aad822-host\") pod \"node-ca-2q7bk\" (UID: \"9cf7e8f8-8031-463c-8267-255938aad822\") " pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.329503 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9cf7e8f8-8031-463c-8267-255938aad822-serviceca\") pod \"node-ca-2q7bk\" (UID: \"9cf7e8f8-8031-463c-8267-255938aad822\") " pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.329519 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq7tx\" (UniqueName: \"kubernetes.io/projected/9cf7e8f8-8031-463c-8267-255938aad822-kube-api-access-cq7tx\") pod \"node-ca-2q7bk\" (UID: \"9cf7e8f8-8031-463c-8267-255938aad822\") " pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.329576 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9cf7e8f8-8031-463c-8267-255938aad822-host\") pod \"node-ca-2q7bk\" (UID: \"9cf7e8f8-8031-463c-8267-255938aad822\") " pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.330568 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9cf7e8f8-8031-463c-8267-255938aad822-serviceca\") pod \"node-ca-2q7bk\" (UID: \"9cf7e8f8-8031-463c-8267-255938aad822\") " pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.343746 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq7tx\" (UniqueName: \"kubernetes.io/projected/9cf7e8f8-8031-463c-8267-255938aad822-kube-api-access-cq7tx\") pod \"node-ca-2q7bk\" (UID: \"9cf7e8f8-8031-463c-8267-255938aad822\") " pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.437388 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.437437 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"b58883a55552790c35d5aef956599fc86b2461ad141bc0859dabdde482938a83"} Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.438616 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerStarted","Data":"89d2a6eebca1b1433017f06bf2d6a9f984deec622aa79642bd771327663431ab"} Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.439843 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8"} Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.440617 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sj5hq" event={"ID":"31197ccb-a3f9-4130-bca4-ca1c3d3ee065","Type":"ContainerStarted","Data":"71174d8f2736d574c9d8f6ebe15ec0dd695306691957dd27fca8961c00a180b0"} Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.453147 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.464456 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.479065 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.482200 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2q7bk" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.491614 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: W1004 04:34:32.492976 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cf7e8f8_8031_463c_8267_255938aad822.slice/crio-fe76e10a28b9a52e2950a644a4c1b8dd8254a7fe0cd07e1ce2ed6c3dfb8acfaa WatchSource:0}: Error finding container fe76e10a28b9a52e2950a644a4c1b8dd8254a7fe0cd07e1ce2ed6c3dfb8acfaa: Status 404 returned error can't find the container with id fe76e10a28b9a52e2950a644a4c1b8dd8254a7fe0cd07e1ce2ed6c3dfb8acfaa Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.517051 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.558558 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.593348 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.633533 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.673913 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.717574 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.755720 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.793136 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.835359 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:32 crc kubenswrapper[4575]: E1004 04:34:32.835575 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:34:36.835545563 +0000 UTC m=+28.164104377 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.836412 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.875320 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.936532 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:32 crc kubenswrapper[4575]: I1004 04:34:32.936643 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:32 crc kubenswrapper[4575]: E1004 04:34:32.936746 4575 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:32 crc kubenswrapper[4575]: E1004 04:34:32.936871 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:36.936844641 +0000 UTC m=+28.265403515 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:32 crc kubenswrapper[4575]: E1004 04:34:32.936776 4575 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:32 crc kubenswrapper[4575]: E1004 04:34:32.936962 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:36.936942274 +0000 UTC m=+28.265501148 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.037983 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.038027 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.038142 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.038158 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.038168 4575 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.038189 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.038234 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.038249 4575 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.038215 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:37.03820232 +0000 UTC m=+28.366761134 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.038350 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:37.038309103 +0000 UTC m=+28.366867917 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.309656 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.309669 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.310048 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:33 crc kubenswrapper[4575]: E1004 04:34:33.310107 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.448756 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f" exitCode=0 Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.448810 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.450842 4575 generic.go:334] "Generic (PLEG): container finished" podID="02c08151-21fd-4ed4-abd7-3ec7e427df35" containerID="9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6" exitCode=0 Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.450896 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerDied","Data":"9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6"} Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.453074 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sj5hq" event={"ID":"31197ccb-a3f9-4130-bca4-ca1c3d3ee065","Type":"ContainerStarted","Data":"adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649"} Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.466887 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2q7bk" event={"ID":"9cf7e8f8-8031-463c-8267-255938aad822","Type":"ContainerStarted","Data":"702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547"} Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.466953 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2q7bk" event={"ID":"9cf7e8f8-8031-463c-8267-255938aad822","Type":"ContainerStarted","Data":"fe76e10a28b9a52e2950a644a4c1b8dd8254a7fe0cd07e1ce2ed6c3dfb8acfaa"} Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.471466 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.492030 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.512308 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.526478 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.540773 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.555457 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.570230 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.585621 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.602313 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.613887 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.627071 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.639103 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.652479 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.666184 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.678710 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.692288 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.702896 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.716201 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.728987 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.738316 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.750694 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.761404 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.798241 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.834657 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.873274 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.914971 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.952666 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:33 crc kubenswrapper[4575]: I1004 04:34:33.994390 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:34 crc kubenswrapper[4575]: I1004 04:34:34.309652 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:34 crc kubenswrapper[4575]: E1004 04:34:34.309801 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:34 crc kubenswrapper[4575]: I1004 04:34:34.471685 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerStarted","Data":"ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.278816 4575 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.280733 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.280767 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.280777 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.280857 4575 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.286716 4575 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.286984 4575 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.287885 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.287916 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.287926 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.287940 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.287949 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: E1004 04:34:35.304134 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.307603 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.307643 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.307655 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.307673 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.307685 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.309060 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.309060 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:35 crc kubenswrapper[4575]: E1004 04:34:35.309176 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:35 crc kubenswrapper[4575]: E1004 04:34:35.309276 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:35 crc kubenswrapper[4575]: E1004 04:34:35.319034 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.321842 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.321875 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.321885 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.321901 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.321912 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: E1004 04:34:35.332847 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.335949 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.335986 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.335995 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.336009 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.336018 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: E1004 04:34:35.347540 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.350985 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.351016 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.351029 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.351045 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.351056 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: E1004 04:34:35.361661 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: E1004 04:34:35.361822 4575 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.363389 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.363420 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.363428 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.363441 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.363450 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.465177 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.465217 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.465227 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.465242 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.465254 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.475722 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.475758 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.486846 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.499435 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.512182 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.522237 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.533540 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.547264 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.558916 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.567058 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.567109 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.567122 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.567140 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.567155 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.570134 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.584947 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.600103 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.612374 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.625327 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.644433 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.666566 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.669090 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.669128 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.669139 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.669153 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.669163 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.771205 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.771238 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.771251 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.771267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.771277 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.873328 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.873362 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.873370 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.873382 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.873390 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.975706 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.975754 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.975766 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.975786 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:35 crc kubenswrapper[4575]: I1004 04:34:35.975796 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:35Z","lastTransitionTime":"2025-10-04T04:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.077987 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.078030 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.078049 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.078074 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.078086 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:36Z","lastTransitionTime":"2025-10-04T04:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.180565 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.180621 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.180630 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.180643 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.180652 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:36Z","lastTransitionTime":"2025-10-04T04:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.283035 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.283345 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.283357 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.283385 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.283395 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:36Z","lastTransitionTime":"2025-10-04T04:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.309621 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:36 crc kubenswrapper[4575]: E1004 04:34:36.309740 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.385826 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.385866 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.385877 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.385894 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.385905 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:36Z","lastTransitionTime":"2025-10-04T04:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.481596 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.481640 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.482974 4575 generic.go:334] "Generic (PLEG): container finished" podID="02c08151-21fd-4ed4-abd7-3ec7e427df35" containerID="ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0" exitCode=0 Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.483019 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerDied","Data":"ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.487987 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.488011 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.488019 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.488034 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.488043 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:36Z","lastTransitionTime":"2025-10-04T04:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.493872 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.506854 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.520259 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.541149 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.562224 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.579699 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.590303 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.590347 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.590357 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.590370 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.590379 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:36Z","lastTransitionTime":"2025-10-04T04:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.591422 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.603345 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.615124 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.627114 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.638080 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.652391 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.666617 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.678595 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.693445 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.693643 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.693756 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.693839 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.693914 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:36Z","lastTransitionTime":"2025-10-04T04:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.796330 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.796358 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.796366 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.796378 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.796387 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:36Z","lastTransitionTime":"2025-10-04T04:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.878118 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:36 crc kubenswrapper[4575]: E1004 04:34:36.878227 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:34:44.878204418 +0000 UTC m=+36.206763242 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.899091 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.899125 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.899135 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.899151 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.899162 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:36Z","lastTransitionTime":"2025-10-04T04:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.978848 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:36 crc kubenswrapper[4575]: I1004 04:34:36.978898 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:36 crc kubenswrapper[4575]: E1004 04:34:36.979013 4575 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:36 crc kubenswrapper[4575]: E1004 04:34:36.979048 4575 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:36 crc kubenswrapper[4575]: E1004 04:34:36.979100 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:44.979081723 +0000 UTC m=+36.307640537 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:36 crc kubenswrapper[4575]: E1004 04:34:36.979118 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:44.979111054 +0000 UTC m=+36.307669878 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.001265 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.001305 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.001317 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.001332 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.001346 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.080069 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.080132 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.080276 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.080298 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.080310 4575 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.080320 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.080390 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.080404 4575 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.080369 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:45.08035266 +0000 UTC m=+36.408911474 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.080514 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:45.080483304 +0000 UTC m=+36.409042118 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.103518 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.103570 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.103628 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.103645 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.103655 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.206339 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.206375 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.206389 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.206428 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.206442 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.308618 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.308660 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.308668 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.308683 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.308693 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.309120 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.309127 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.309264 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:37 crc kubenswrapper[4575]: E1004 04:34:37.309353 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.410566 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.410630 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.410642 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.410659 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.410668 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.487796 4575 generic.go:334] "Generic (PLEG): container finished" podID="02c08151-21fd-4ed4-abd7-3ec7e427df35" containerID="ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c" exitCode=0 Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.487866 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerDied","Data":"ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.492024 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.492063 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.504256 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.513260 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.513299 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.513310 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.513326 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.513338 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.520140 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.534858 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.551728 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.564374 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.575253 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.587110 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.598555 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.609648 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.615050 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.615102 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.615116 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.615135 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.615156 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.619417 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.629254 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.642178 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.653470 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.666143 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:37Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.723788 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.723823 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.723835 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.723851 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.723868 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.826917 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.826964 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.826976 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.826993 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.827008 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.929665 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.929952 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.930051 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.930140 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:37 crc kubenswrapper[4575]: I1004 04:34:37.930224 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:37Z","lastTransitionTime":"2025-10-04T04:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.032992 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.033028 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.033039 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.033057 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.033068 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.135126 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.135167 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.135183 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.135197 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.135207 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.237365 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.237396 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.237406 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.237422 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.237430 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.309321 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:38 crc kubenswrapper[4575]: E1004 04:34:38.309475 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.339897 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.339935 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.339943 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.339958 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.339968 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.443098 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.443138 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.443151 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.443174 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.443185 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.496868 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerStarted","Data":"5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.545376 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.545663 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.545747 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.545924 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.546015 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.648876 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.648913 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.648924 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.648940 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.648953 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.751738 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.751775 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.751784 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.751798 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.751807 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.854694 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.854727 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.854735 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.854748 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.854783 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.956303 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.956350 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.956366 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.956389 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:38 crc kubenswrapper[4575]: I1004 04:34:38.956405 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:38Z","lastTransitionTime":"2025-10-04T04:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.061641 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.061681 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.061692 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.061708 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.061721 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.163839 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.163885 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.163894 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.163911 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.163924 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.266439 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.266474 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.266482 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.266496 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.266505 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.309704 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.309749 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:39 crc kubenswrapper[4575]: E1004 04:34:39.309812 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:39 crc kubenswrapper[4575]: E1004 04:34:39.309883 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.321797 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.337129 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.348020 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.357873 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.367076 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.368509 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.368547 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.368557 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.368572 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.368596 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.383410 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.393983 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.405116 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.416066 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.429346 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.441010 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.454346 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.464050 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.473306 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.473371 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.473385 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.473402 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.473488 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.477253 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.501349 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.503405 4575 generic.go:334] "Generic (PLEG): container finished" podID="02c08151-21fd-4ed4-abd7-3ec7e427df35" containerID="5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb" exitCode=0 Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.503511 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerDied","Data":"5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.521181 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.549065 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.579019 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.579051 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.579059 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.579073 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.579082 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.586369 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.610139 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.621141 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.635014 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.646525 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.666601 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.679000 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.680545 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.680572 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.680581 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.680608 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.680617 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.690903 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.701714 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.713781 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.725416 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.735438 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.783123 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.783424 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.783435 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.783456 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.783469 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.886070 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.886114 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.886122 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.886147 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.886156 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.989380 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.989454 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.989477 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.989505 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:39 crc kubenswrapper[4575]: I1004 04:34:39.989531 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:39Z","lastTransitionTime":"2025-10-04T04:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.092606 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.092667 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.092689 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.092711 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.092725 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:40Z","lastTransitionTime":"2025-10-04T04:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.195628 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.195685 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.195706 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.195727 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.195741 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:40Z","lastTransitionTime":"2025-10-04T04:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.298408 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.298464 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.298479 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.298497 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.298511 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:40Z","lastTransitionTime":"2025-10-04T04:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.309718 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:40 crc kubenswrapper[4575]: E1004 04:34:40.309875 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.401234 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.401276 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.401285 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.401299 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.401307 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:40Z","lastTransitionTime":"2025-10-04T04:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.504727 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.504772 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.504781 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.504795 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.504808 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:40Z","lastTransitionTime":"2025-10-04T04:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.515497 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerStarted","Data":"ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.529694 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.538969 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.551013 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.561407 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.576790 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.588615 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.600252 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.607389 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.607442 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.607453 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.607470 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.607481 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:40Z","lastTransitionTime":"2025-10-04T04:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.612412 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.623296 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.635703 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.648435 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.661937 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.671383 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.684825 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:40Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.710869 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.710909 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.710918 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.710932 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.710943 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:40Z","lastTransitionTime":"2025-10-04T04:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.812972 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.813005 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.813013 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.813028 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.813037 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:40Z","lastTransitionTime":"2025-10-04T04:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.915194 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.915230 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.915241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.915255 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:40 crc kubenswrapper[4575]: I1004 04:34:40.915264 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:40Z","lastTransitionTime":"2025-10-04T04:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.017568 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.017634 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.017668 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.017685 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.017699 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.119147 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.119179 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.119187 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.119200 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.119209 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.200993 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.219652 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.221159 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.221200 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.221208 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.221225 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.221235 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.235666 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.248438 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.258817 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.269095 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.283788 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.303022 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.309193 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.309234 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:41 crc kubenswrapper[4575]: E1004 04:34:41.309470 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:41 crc kubenswrapper[4575]: E1004 04:34:41.309351 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.314430 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.322998 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.323033 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.323042 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.323055 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.323083 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.329455 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.344412 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.357309 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.373391 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.385007 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.396632 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.425255 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.425490 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.425618 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.425713 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.425817 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.520883 4575 generic.go:334] "Generic (PLEG): container finished" podID="02c08151-21fd-4ed4-abd7-3ec7e427df35" containerID="ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83" exitCode=0 Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.520946 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerDied","Data":"ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.527410 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.527437 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.527448 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.527463 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.527473 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.528266 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.528778 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.528807 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.537152 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.548435 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.562045 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.563052 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.563202 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.578529 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.596853 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.609761 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.625872 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.630273 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.630301 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.630312 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.630330 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.630340 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.637038 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.650750 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.662862 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.674724 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.686010 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.702490 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.715133 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.731317 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.732974 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.733028 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.733039 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.733060 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.733075 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.744410 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.762951 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.774766 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.786773 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.800762 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.813848 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.826981 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.835306 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.835338 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.835345 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.835362 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.835370 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.839573 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.859342 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.869517 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.883386 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.898277 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.913111 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:41Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.937746 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.937813 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.937827 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.937850 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:41 crc kubenswrapper[4575]: I1004 04:34:41.937866 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:41Z","lastTransitionTime":"2025-10-04T04:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.040997 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.041049 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.041064 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.041083 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.041106 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.144627 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.144691 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.144710 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.144738 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.144758 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.247969 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.248003 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.248012 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.248026 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.248038 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.309330 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:42 crc kubenswrapper[4575]: E1004 04:34:42.309484 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.350343 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.350619 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.350629 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.350642 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.350651 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.452894 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.453171 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.453245 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.453321 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.453385 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.535094 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.536349 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerStarted","Data":"9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.549277 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.561384 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.561426 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.561441 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.561461 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.561480 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.569417 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.583154 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.595853 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.605320 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.618997 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.633990 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.651263 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.664206 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.664240 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.664248 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.664261 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.664270 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.668485 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.682294 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.692481 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.702784 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.713573 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.723818 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.763949 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7"] Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.764395 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.766018 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.766539 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.766567 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.766577 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.766610 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.766622 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.766732 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.777900 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.788438 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.796812 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.808921 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.817129 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.828793 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.838805 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.838853 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.838883 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f6l6\" (UniqueName: \"kubernetes.io/projected/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-kube-api-access-5f6l6\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.838915 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.839009 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.848215 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.858319 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.871132 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.871169 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.871179 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.871196 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.871208 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.883464 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.897490 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.907866 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.918030 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.927830 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.938295 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.939684 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.939728 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.939757 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.939795 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f6l6\" (UniqueName: \"kubernetes.io/projected/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-kube-api-access-5f6l6\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.940486 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.941118 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.945413 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.954143 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f6l6\" (UniqueName: \"kubernetes.io/projected/331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1-kube-api-access-5f6l6\") pod \"ovnkube-control-plane-749d76644c-fqfm7\" (UID: \"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.974105 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.974143 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.974155 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.974173 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:42 crc kubenswrapper[4575]: I1004 04:34:42.974184 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:42Z","lastTransitionTime":"2025-10-04T04:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.076298 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.076342 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.076356 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.076370 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.076380 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:43Z","lastTransitionTime":"2025-10-04T04:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.178625 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.178666 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.178693 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.178709 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.178720 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:43Z","lastTransitionTime":"2025-10-04T04:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.198847 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" Oct 04 04:34:43 crc kubenswrapper[4575]: W1004 04:34:43.211209 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod331e96fa_b5b5_484d_a2d7_bdbfa6da8ea1.slice/crio-83cad5a73d7adaf05c34514333ef46722f451fd7ebad8d44a3c7abbb1b304ab8 WatchSource:0}: Error finding container 83cad5a73d7adaf05c34514333ef46722f451fd7ebad8d44a3c7abbb1b304ab8: Status 404 returned error can't find the container with id 83cad5a73d7adaf05c34514333ef46722f451fd7ebad8d44a3c7abbb1b304ab8 Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.281386 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.281433 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.281445 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.281467 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.281479 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:43Z","lastTransitionTime":"2025-10-04T04:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.309214 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.309273 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:43 crc kubenswrapper[4575]: E1004 04:34:43.309389 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:43 crc kubenswrapper[4575]: E1004 04:34:43.309507 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.383607 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.383635 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.383644 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.383657 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.383666 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:43Z","lastTransitionTime":"2025-10-04T04:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.487394 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.487444 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.487481 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.487499 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.487509 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:43Z","lastTransitionTime":"2025-10-04T04:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.539032 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" event={"ID":"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1","Type":"ContainerStarted","Data":"83cad5a73d7adaf05c34514333ef46722f451fd7ebad8d44a3c7abbb1b304ab8"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.542198 4575 generic.go:334] "Generic (PLEG): container finished" podID="02c08151-21fd-4ed4-abd7-3ec7e427df35" containerID="9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a" exitCode=0 Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.542264 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerDied","Data":"9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.542365 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.556761 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.568341 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.580863 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.589869 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.589911 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.589932 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.589949 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.589960 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:43Z","lastTransitionTime":"2025-10-04T04:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.591970 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.601928 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.612545 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.625524 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.639767 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.649547 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.662538 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.673042 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.680840 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.691725 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.692838 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.692869 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.692879 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.692894 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.692903 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:43Z","lastTransitionTime":"2025-10-04T04:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.704385 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.725392 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.795329 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.795375 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.795387 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.795402 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.795413 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:43Z","lastTransitionTime":"2025-10-04T04:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.838791 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9nvms"] Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.839417 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:43 crc kubenswrapper[4575]: E1004 04:34:43.839539 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.854049 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.863651 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.875791 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.889228 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.898978 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.899016 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.899025 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.899040 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.899053 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:43Z","lastTransitionTime":"2025-10-04T04:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.905545 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.919336 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.932400 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.943430 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.948709 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ct29\" (UniqueName: \"kubernetes.io/projected/026aaa63-845a-45bb-bc23-7b7c3024eead-kube-api-access-9ct29\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.948810 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.955065 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.968061 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.979874 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:43 crc kubenswrapper[4575]: I1004 04:34:43.992527 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.001288 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.001323 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.001331 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.001345 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.001354 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.006084 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.016982 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.031221 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.042465 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.050004 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ct29\" (UniqueName: \"kubernetes.io/projected/026aaa63-845a-45bb-bc23-7b7c3024eead-kube-api-access-9ct29\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.050095 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:44 crc kubenswrapper[4575]: E1004 04:34:44.050197 4575 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:44 crc kubenswrapper[4575]: E1004 04:34:44.050260 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs podName:026aaa63-845a-45bb-bc23-7b7c3024eead nodeName:}" failed. No retries permitted until 2025-10-04 04:34:44.550239715 +0000 UTC m=+35.878798549 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs") pod "network-metrics-daemon-9nvms" (UID: "026aaa63-845a-45bb-bc23-7b7c3024eead") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.065233 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ct29\" (UniqueName: \"kubernetes.io/projected/026aaa63-845a-45bb-bc23-7b7c3024eead-kube-api-access-9ct29\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.103704 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.103747 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.103756 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.103770 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.103779 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.206005 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.206037 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.206046 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.206058 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.206069 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.309234 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.309257 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.309266 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.309280 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.309289 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.309301 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:44 crc kubenswrapper[4575]: E1004 04:34:44.309402 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.411558 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.411622 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.411635 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.411652 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.411663 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.514191 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.514227 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.514236 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.514249 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.514258 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.545933 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" event={"ID":"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1","Type":"ContainerStarted","Data":"a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.549022 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" event={"ID":"02c08151-21fd-4ed4-abd7-3ec7e427df35","Type":"ContainerStarted","Data":"ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.554642 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:44 crc kubenswrapper[4575]: E1004 04:34:44.554785 4575 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:44 crc kubenswrapper[4575]: E1004 04:34:44.554839 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs podName:026aaa63-845a-45bb-bc23-7b7c3024eead nodeName:}" failed. No retries permitted until 2025-10-04 04:34:45.554826227 +0000 UTC m=+36.883385041 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs") pod "network-metrics-daemon-9nvms" (UID: "026aaa63-845a-45bb-bc23-7b7c3024eead") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.617032 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.617076 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.617087 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.617104 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.617115 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.719973 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.720301 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.720310 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.720325 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.720334 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.823623 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.823698 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.823718 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.823751 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.823772 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.926870 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.926921 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.926933 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.926953 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.926968 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:44Z","lastTransitionTime":"2025-10-04T04:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:44 crc kubenswrapper[4575]: I1004 04:34:44.959324 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:44 crc kubenswrapper[4575]: E1004 04:34:44.959504 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:00.959472649 +0000 UTC m=+52.288031463 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.030936 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.031009 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.031023 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.031043 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.031059 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.060185 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.060232 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.060375 4575 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.060434 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:01.060420447 +0000 UTC m=+52.388979251 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.060499 4575 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.060637 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:01.060581441 +0000 UTC m=+52.389140295 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.134708 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.134749 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.134761 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.134776 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.134788 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.161641 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.161720 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.161817 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.161832 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.161841 4575 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.161882 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:01.161865428 +0000 UTC m=+52.490424242 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.162085 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.162155 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.162179 4575 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.162299 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:01.16226614 +0000 UTC m=+52.490824984 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.237570 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.237656 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.237675 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.237698 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.237713 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.309030 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.309074 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.309030 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.309153 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.309243 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.309312 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.341536 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.341563 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.341572 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.341602 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.341611 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.444459 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.444496 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.444507 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.444521 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.444530 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.546980 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.547027 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.547040 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.547059 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.547072 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.552647 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" event={"ID":"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1","Type":"ContainerStarted","Data":"222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.566415 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.566940 4575 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.567067 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs podName:026aaa63-845a-45bb-bc23-7b7c3024eead nodeName:}" failed. No retries permitted until 2025-10-04 04:34:47.567039887 +0000 UTC m=+38.895598751 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs") pod "network-metrics-daemon-9nvms" (UID: "026aaa63-845a-45bb-bc23-7b7c3024eead") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.573603 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.586538 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.601434 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.613695 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.627853 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.639484 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.642383 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.642419 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.642427 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.642442 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.642452 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.655108 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.657782 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.658999 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.659096 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.659187 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.659275 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.659351 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.669579 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.672173 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.676705 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.676769 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.676783 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.676809 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.676824 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.680600 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.688031 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.692399 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.692442 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.692454 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.692475 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.692489 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.693192 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.706117 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.709639 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.709746 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.709815 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.709886 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.709943 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.709868 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.736553 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: E1004 04:34:45.736735 4575 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.738652 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.738694 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.738706 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.738722 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.738733 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.739140 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.755646 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.770616 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.805721 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.824515 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.841156 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.841201 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.841213 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.841229 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.841240 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.943385 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.943445 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.943456 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.943471 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:45 crc kubenswrapper[4575]: I1004 04:34:45.943482 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:45Z","lastTransitionTime":"2025-10-04T04:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.046072 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.046426 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.046499 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.046599 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.046672 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.149389 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.149436 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.149449 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.149468 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.149484 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.251097 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.251126 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.251133 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.251146 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.251154 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.309253 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:46 crc kubenswrapper[4575]: E1004 04:34:46.309391 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.353266 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.353324 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.353333 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.353346 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.353356 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.456178 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.456217 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.456227 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.456240 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.456249 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.557936 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.557987 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.558003 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.558024 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.558039 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.574147 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.592213 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.602795 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.616947 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.627190 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.636909 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.648271 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.656592 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.660085 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.660120 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.660129 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.660143 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.660152 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.668868 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.681141 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.701675 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.713144 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.723606 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.734891 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.745668 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.759025 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.762679 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.762714 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.762726 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.762741 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.762751 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.865139 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.865174 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.865184 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.865197 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.865206 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.878418 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.878574 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.892078 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" probeResult="failure" output="" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.903065 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" probeResult="failure" output="" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.967655 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.967694 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.967704 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.967719 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:46 crc kubenswrapper[4575]: I1004 04:34:46.967731 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:46Z","lastTransitionTime":"2025-10-04T04:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.069745 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.069773 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.069781 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.069793 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.069800 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:47Z","lastTransitionTime":"2025-10-04T04:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.172315 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.172349 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.172360 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.172377 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.172388 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:47Z","lastTransitionTime":"2025-10-04T04:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.274723 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.274766 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.274775 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.274819 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.274829 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:47Z","lastTransitionTime":"2025-10-04T04:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.309451 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.309526 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.309551 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:47 crc kubenswrapper[4575]: E1004 04:34:47.309702 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:47 crc kubenswrapper[4575]: E1004 04:34:47.309755 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:47 crc kubenswrapper[4575]: E1004 04:34:47.309853 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.377422 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.377465 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.377477 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.377494 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.377504 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:47Z","lastTransitionTime":"2025-10-04T04:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.479807 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.479832 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.479840 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.479852 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.479860 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:47Z","lastTransitionTime":"2025-10-04T04:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.559464 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/0.log" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.562158 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4" exitCode=1 Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.562190 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.562847 4575 scope.go:117] "RemoveContainer" containerID="cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.576051 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.581364 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.581390 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.581399 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.581411 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.581420 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:47Z","lastTransitionTime":"2025-10-04T04:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.586093 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.586217 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:47 crc kubenswrapper[4575]: E1004 04:34:47.586361 4575 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:47 crc kubenswrapper[4575]: E1004 04:34:47.586406 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs podName:026aaa63-845a-45bb-bc23-7b7c3024eead nodeName:}" failed. No retries permitted until 2025-10-04 04:34:51.586393543 +0000 UTC m=+42.914952357 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs") pod "network-metrics-daemon-9nvms" (UID: "026aaa63-845a-45bb-bc23-7b7c3024eead") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.597374 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.608566 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.626257 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:47Z\\\",\\\"message\\\":\\\"annel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1004 04:34:46.969520 5792 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 04:34:46.969542 5792 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:34:46.969615 5792 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1004 04:34:46.970045 5792 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 04:34:46.970060 5792 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 04:34:46.970067 5792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:34:46.970121 5792 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 04:34:46.970147 5792 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 04:34:46.970171 5792 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:34:46.970179 5792 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 04:34:46.970259 5792 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 04:34:46.970342 5792 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 04:34:46.970412 5792 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 04:34:46.970503 5792 factory.go:656] Stopping watch factory\\\\nI1004 04:34:46.970515 5792 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:34:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.642373 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.657106 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.674394 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.683446 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.683629 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.683718 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.683952 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.684039 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:47Z","lastTransitionTime":"2025-10-04T04:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.685261 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.698176 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.713735 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.726364 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.740692 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.752519 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.761593 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.775670 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.786713 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.786762 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.786774 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.786790 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.786802 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:47Z","lastTransitionTime":"2025-10-04T04:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.921533 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.921557 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.921566 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.921579 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:47 crc kubenswrapper[4575]: I1004 04:34:47.921606 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:47Z","lastTransitionTime":"2025-10-04T04:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.023489 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.023779 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.023868 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.023954 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.024059 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.126533 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.126558 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.126566 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.126578 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.126597 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.228701 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.228760 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.228770 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.228784 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.228794 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.309973 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:48 crc kubenswrapper[4575]: E1004 04:34:48.310211 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.331966 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.332031 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.332049 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.332075 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.332093 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.434437 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.434473 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.434483 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.434498 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.434509 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.536802 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.536834 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.536842 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.536853 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.536864 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.566696 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/0.log" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.570559 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.571014 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.585286 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.598697 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.611857 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.623296 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.636911 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.638756 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.638795 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.638804 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.638819 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.638829 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.651608 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.663774 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.679651 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.692444 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.710014 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:47Z\\\",\\\"message\\\":\\\"annel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1004 04:34:46.969520 5792 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 04:34:46.969542 5792 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:34:46.969615 5792 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1004 04:34:46.970045 5792 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 04:34:46.970060 5792 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 04:34:46.970067 5792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:34:46.970121 5792 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 04:34:46.970147 5792 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 04:34:46.970171 5792 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:34:46.970179 5792 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 04:34:46.970259 5792 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 04:34:46.970342 5792 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 04:34:46.970412 5792 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 04:34:46.970503 5792 factory.go:656] Stopping watch factory\\\\nI1004 04:34:46.970515 5792 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:34:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.722793 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.733835 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.740828 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.740864 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.740878 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.740895 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.740906 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.745749 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.757317 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.769348 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.780349 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.843437 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.843483 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.843494 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.843509 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.843521 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.945693 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.945792 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.945808 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.945823 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:48 crc kubenswrapper[4575]: I1004 04:34:48.945834 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:48Z","lastTransitionTime":"2025-10-04T04:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.047532 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.047799 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.047862 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.047920 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.047982 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.150044 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.150279 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.150343 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.150404 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.150509 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.252981 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.253021 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.253030 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.253044 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.253054 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.309172 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:49 crc kubenswrapper[4575]: E1004 04:34:49.309347 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.309185 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:49 crc kubenswrapper[4575]: E1004 04:34:49.309504 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.309569 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:49 crc kubenswrapper[4575]: E1004 04:34:49.309649 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.324790 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.334635 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.347347 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.355109 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.355143 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.355155 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.355172 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.355183 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.359837 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.382243 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:47Z\\\",\\\"message\\\":\\\"annel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1004 04:34:46.969520 5792 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 04:34:46.969542 5792 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:34:46.969615 5792 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1004 04:34:46.970045 5792 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 04:34:46.970060 5792 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 04:34:46.970067 5792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:34:46.970121 5792 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 04:34:46.970147 5792 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 04:34:46.970171 5792 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:34:46.970179 5792 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 04:34:46.970259 5792 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 04:34:46.970342 5792 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 04:34:46.970412 5792 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 04:34:46.970503 5792 factory.go:656] Stopping watch factory\\\\nI1004 04:34:46.970515 5792 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:34:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.393233 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.403851 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.413972 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.429457 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.456969 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.457192 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.457267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.457351 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.457429 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.460191 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.472231 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.483506 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.496686 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.505476 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.513938 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.524477 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.559237 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.559267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.559276 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.559288 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.559297 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.574055 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/1.log" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.574927 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/0.log" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.577619 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a" exitCode=1 Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.577654 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.577700 4575 scope.go:117] "RemoveContainer" containerID="cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.578726 4575 scope.go:117] "RemoveContainer" containerID="0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a" Oct 04 04:34:49 crc kubenswrapper[4575]: E1004 04:34:49.578977 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.597857 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.607827 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.619218 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.639391 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.661128 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.661165 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.661173 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.661187 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.661196 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.663078 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbaedb35cad18bb79deea00e3f85ffa46eb91377201151d857c4cc699bec2d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:47Z\\\",\\\"message\\\":\\\"annel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1004 04:34:46.969520 5792 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 04:34:46.969542 5792 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:34:46.969615 5792 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1004 04:34:46.970045 5792 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 04:34:46.970060 5792 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 04:34:46.970067 5792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:34:46.970121 5792 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 04:34:46.970147 5792 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 04:34:46.970171 5792 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:34:46.970179 5792 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 04:34:46.970259 5792 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 04:34:46.970342 5792 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 04:34:46.970412 5792 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 04:34:46.970503 5792 factory.go:656] Stopping watch factory\\\\nI1004 04:34:46.970515 5792 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:34:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:48Z\\\",\\\"message\\\":\\\"ere:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:34:48.863920 6032 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:34:48.863926 6032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.677843 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.689845 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.701948 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.712680 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.723914 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.735789 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.749784 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.759917 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.763113 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.763140 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.763150 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.763163 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.763173 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.775685 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.787405 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.800811 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:49Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.864960 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.864994 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.865024 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.865039 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.865049 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.967232 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.967269 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.967281 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.967296 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:49 crc kubenswrapper[4575]: I1004 04:34:49.967309 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:49Z","lastTransitionTime":"2025-10-04T04:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.069180 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.069436 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.069566 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.069699 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.069759 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.172141 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.172184 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.172195 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.172213 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.172226 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.274550 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.274876 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.274944 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.275036 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.275104 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.308941 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:50 crc kubenswrapper[4575]: E1004 04:34:50.309072 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.377677 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.377729 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.377739 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.377754 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.377767 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.480475 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.480509 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.480518 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.480533 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.480542 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.582404 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.582480 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.582491 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.582505 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.582514 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.582823 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/1.log" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.586234 4575 scope.go:117] "RemoveContainer" containerID="0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a" Oct 04 04:34:50 crc kubenswrapper[4575]: E1004 04:34:50.586366 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.604448 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.619321 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.641567 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.658830 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.672324 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.684272 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.684310 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.684319 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.684333 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.684343 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.685266 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.698456 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.709345 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.722205 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.734325 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.755284 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:48Z\\\",\\\"message\\\":\\\"ere:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:34:48.863920 6032 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:34:48.863926 6032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.768749 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.781774 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.786628 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.786890 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.786979 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.787056 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.787135 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.794479 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.806627 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.821760 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:50Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.889876 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.890175 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.890310 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.890377 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.890438 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.992350 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.992387 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.992397 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.992412 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:50 crc kubenswrapper[4575]: I1004 04:34:50.992424 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:50Z","lastTransitionTime":"2025-10-04T04:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.096226 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.096270 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.096282 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.096299 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.096311 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:51Z","lastTransitionTime":"2025-10-04T04:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.198712 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.198745 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.198753 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.198770 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.198781 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:51Z","lastTransitionTime":"2025-10-04T04:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.301078 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.301336 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.301416 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.301496 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.301573 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:51Z","lastTransitionTime":"2025-10-04T04:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.309446 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.309485 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:51 crc kubenswrapper[4575]: E1004 04:34:51.309562 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.309451 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:51 crc kubenswrapper[4575]: E1004 04:34:51.309677 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:51 crc kubenswrapper[4575]: E1004 04:34:51.309764 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.403916 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.403951 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.403962 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.403976 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.403987 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:51Z","lastTransitionTime":"2025-10-04T04:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.506698 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.506742 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.506751 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.506769 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.506778 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:51Z","lastTransitionTime":"2025-10-04T04:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.608982 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.609049 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.609062 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.609078 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.609089 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:51Z","lastTransitionTime":"2025-10-04T04:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.651043 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:51 crc kubenswrapper[4575]: E1004 04:34:51.651161 4575 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:51 crc kubenswrapper[4575]: E1004 04:34:51.651210 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs podName:026aaa63-845a-45bb-bc23-7b7c3024eead nodeName:}" failed. No retries permitted until 2025-10-04 04:34:59.651193972 +0000 UTC m=+50.979752786 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs") pod "network-metrics-daemon-9nvms" (UID: "026aaa63-845a-45bb-bc23-7b7c3024eead") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.711319 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.711515 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.711616 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.711729 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.711800 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:51Z","lastTransitionTime":"2025-10-04T04:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.814352 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.814403 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.814414 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.814429 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.814440 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:51Z","lastTransitionTime":"2025-10-04T04:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.917003 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.917030 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.917038 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.917050 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:51 crc kubenswrapper[4575]: I1004 04:34:51.917058 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:51Z","lastTransitionTime":"2025-10-04T04:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.019046 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.019276 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.019503 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.019682 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.019843 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.123107 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.123351 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.123361 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.123377 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.123389 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.225189 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.225223 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.225231 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.225245 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.225253 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.309212 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:52 crc kubenswrapper[4575]: E1004 04:34:52.309345 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.327743 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.327792 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.327803 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.327819 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.327829 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.430636 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.430670 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.430681 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.430694 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.430703 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.532888 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.532917 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.532925 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.532938 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.532946 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.635286 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.635816 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.635881 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.635945 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.636002 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.740488 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.740537 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.740546 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.740608 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.740618 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.842890 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.842956 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.842968 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.842985 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.843003 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.945408 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.945437 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.945463 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.945479 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:52 crc kubenswrapper[4575]: I1004 04:34:52.945488 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:52Z","lastTransitionTime":"2025-10-04T04:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.047747 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.047799 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.047813 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.047863 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.047878 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.150115 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.150414 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.150532 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.150679 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.150883 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.253174 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.253575 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.253765 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.254110 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.254291 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.309346 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.309372 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.309395 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:53 crc kubenswrapper[4575]: E1004 04:34:53.310262 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:53 crc kubenswrapper[4575]: E1004 04:34:53.310338 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:34:53 crc kubenswrapper[4575]: E1004 04:34:53.310772 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.356917 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.356948 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.356956 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.356971 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.356981 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.459617 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.459661 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.459696 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.459718 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.459737 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.561902 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.561933 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.561941 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.561953 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.561963 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.665035 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.665546 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.665644 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.665722 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.665781 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.769336 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.769634 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.769712 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.769784 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.769844 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.872089 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.872155 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.872173 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.872195 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.872211 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.974794 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.975069 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.975159 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.975262 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:53 crc kubenswrapper[4575]: I1004 04:34:53.975432 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:53Z","lastTransitionTime":"2025-10-04T04:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.077478 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.077516 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.077528 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.077544 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.077555 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:54Z","lastTransitionTime":"2025-10-04T04:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.179408 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.179442 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.179453 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.179468 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.179479 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:54Z","lastTransitionTime":"2025-10-04T04:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.283019 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.283286 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.283373 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.283472 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.283561 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:54Z","lastTransitionTime":"2025-10-04T04:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.309729 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:54 crc kubenswrapper[4575]: E1004 04:34:54.310329 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.385616 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.385644 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.385651 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.385663 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.385671 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:54Z","lastTransitionTime":"2025-10-04T04:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.488474 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.489022 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.489211 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.489389 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.489564 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:54Z","lastTransitionTime":"2025-10-04T04:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.591796 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.591836 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.591859 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.591878 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.591888 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:54Z","lastTransitionTime":"2025-10-04T04:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.694070 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.694139 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.694151 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.694167 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.694177 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:54Z","lastTransitionTime":"2025-10-04T04:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.797884 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.797949 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.797966 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.797988 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.798005 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:54Z","lastTransitionTime":"2025-10-04T04:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.899836 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.899912 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.899937 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.899965 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:54 crc kubenswrapper[4575]: I1004 04:34:54.899985 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:54Z","lastTransitionTime":"2025-10-04T04:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.002622 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.002681 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.002692 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.002707 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.002720 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.104729 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.104971 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.105038 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.105107 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.105167 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.207887 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.208184 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.208292 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.208407 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.208496 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.309396 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.309765 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.309987 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:55 crc kubenswrapper[4575]: E1004 04:34:55.310180 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:55 crc kubenswrapper[4575]: E1004 04:34:55.310794 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:34:55 crc kubenswrapper[4575]: E1004 04:34:55.310916 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.311058 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.311250 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.311403 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.311528 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.311695 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.414543 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.414839 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.414932 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.415038 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.415108 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.518016 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.518056 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.518065 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.518081 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.518093 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.620495 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.620527 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.620536 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.620550 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.620561 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.723329 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.723380 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.723396 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.723418 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.723435 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.791781 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.791824 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.791837 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.791855 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.791866 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: E1004 04:34:55.807115 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.811362 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.811403 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.811434 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.811457 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.811468 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: E1004 04:34:55.824100 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.827313 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.827353 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.827361 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.827375 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.827383 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: E1004 04:34:55.837833 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.841095 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.841208 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.841268 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.841347 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.841539 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: E1004 04:34:55.854742 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.858514 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.858551 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.858561 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.858581 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.858622 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: E1004 04:34:55.868920 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:55 crc kubenswrapper[4575]: E1004 04:34:55.869047 4575 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.870306 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.870416 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.870490 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.870562 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.870665 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.973370 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.973422 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.973434 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.973449 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:55 crc kubenswrapper[4575]: I1004 04:34:55.973461 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:55Z","lastTransitionTime":"2025-10-04T04:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.075733 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.075769 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.075780 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.075795 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.075805 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:56Z","lastTransitionTime":"2025-10-04T04:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.177918 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.178166 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.178233 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.178315 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.178384 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:56Z","lastTransitionTime":"2025-10-04T04:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.281043 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.281094 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.281108 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.281124 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.281132 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:56Z","lastTransitionTime":"2025-10-04T04:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.308962 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:56 crc kubenswrapper[4575]: E1004 04:34:56.309122 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.384774 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.384874 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.384898 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.384930 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.384952 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:56Z","lastTransitionTime":"2025-10-04T04:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.487705 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.488025 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.488160 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.488280 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.488421 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:56Z","lastTransitionTime":"2025-10-04T04:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.591016 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.591280 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.591347 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.591422 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.591486 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:56Z","lastTransitionTime":"2025-10-04T04:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.694603 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.694651 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.694666 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.694687 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.694700 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:56Z","lastTransitionTime":"2025-10-04T04:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.796845 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.797337 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.797427 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.797486 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.797552 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:56Z","lastTransitionTime":"2025-10-04T04:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.900073 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.900452 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.900574 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.900699 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:56 crc kubenswrapper[4575]: I1004 04:34:56.900786 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:56Z","lastTransitionTime":"2025-10-04T04:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.003061 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.003259 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.003340 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.003402 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.003456 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.105218 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.105501 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.105622 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.105698 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.105766 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.207436 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.207517 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.207528 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.207545 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.207555 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.308917 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.308965 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:57 crc kubenswrapper[4575]: E1004 04:34:57.309321 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.309014 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:57 crc kubenswrapper[4575]: E1004 04:34:57.309255 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:34:57 crc kubenswrapper[4575]: E1004 04:34:57.309417 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.309931 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.310017 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.310028 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.310040 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.310050 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.412088 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.412119 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.412127 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.412144 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.412154 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.514018 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.514051 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.514060 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.514074 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.514084 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.615312 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.615351 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.615361 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.615378 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.615389 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.717202 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.717230 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.717238 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.717251 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.717259 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.819568 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.819705 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.819717 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.819729 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.819738 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.921467 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.921509 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.921526 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.921542 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:57 crc kubenswrapper[4575]: I1004 04:34:57.921553 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:57Z","lastTransitionTime":"2025-10-04T04:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.023750 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.023774 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.023782 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.023795 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.023803 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.125835 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.125873 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.125884 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.125898 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.125911 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.128298 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.137910 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.142153 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.153937 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.166223 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.178179 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.189167 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.200707 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.214002 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.224334 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.227841 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.227863 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.227874 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.227891 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.227902 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.242503 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.254254 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.275891 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.291837 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.301166 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.308820 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:58 crc kubenswrapper[4575]: E1004 04:34:58.308928 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.312574 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.323474 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.331059 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.331083 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.331093 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.331119 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.331128 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.340395 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:48Z\\\",\\\"message\\\":\\\"ere:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:34:48.863920 6032 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:34:48.863926 6032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.432878 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.432921 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.432937 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.432957 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.432972 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.535785 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.535836 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.535851 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.535870 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.535885 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.638408 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.638467 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.638484 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.638507 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.638524 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.740741 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.740777 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.740789 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.740806 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.740818 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.842894 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.842926 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.842934 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.842946 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.842954 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.944958 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.945001 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.945010 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.945023 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:58 crc kubenswrapper[4575]: I1004 04:34:58.945032 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:58Z","lastTransitionTime":"2025-10-04T04:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.047047 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.047078 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.047088 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.047114 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.047125 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.149575 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.149632 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.149641 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.149655 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.149667 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.251818 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.251848 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.251856 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.251868 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.251877 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.309697 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.309706 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.309814 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:59 crc kubenswrapper[4575]: E1004 04:34:59.309915 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:59 crc kubenswrapper[4575]: E1004 04:34:59.309976 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:34:59 crc kubenswrapper[4575]: E1004 04:34:59.310113 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.324148 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.336095 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.354563 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.354627 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.354640 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.354659 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.354672 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.356899 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:48Z\\\",\\\"message\\\":\\\"ere:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:34:48.863920 6032 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:34:48.863926 6032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.369031 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.391665 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.401433 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.413614 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.426755 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.440352 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.450367 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.456770 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.456824 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.456839 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.456860 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.456874 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.466193 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.480556 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.494528 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.510951 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.523343 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.536745 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.547694 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.559333 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.559376 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.559387 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.559402 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.559413 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.661673 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.661725 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.661748 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.661770 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.661788 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.734083 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:34:59 crc kubenswrapper[4575]: E1004 04:34:59.734224 4575 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:59 crc kubenswrapper[4575]: E1004 04:34:59.734278 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs podName:026aaa63-845a-45bb-bc23-7b7c3024eead nodeName:}" failed. No retries permitted until 2025-10-04 04:35:15.734264775 +0000 UTC m=+67.062823589 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs") pod "network-metrics-daemon-9nvms" (UID: "026aaa63-845a-45bb-bc23-7b7c3024eead") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.764986 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.765019 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.765029 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.765041 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.765050 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.867616 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.867653 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.867663 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.867679 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.867691 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.969624 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.969657 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.969664 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.969676 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:59 crc kubenswrapper[4575]: I1004 04:34:59.969685 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:59Z","lastTransitionTime":"2025-10-04T04:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.072112 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.072147 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.072155 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.072167 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.072177 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.174760 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.175053 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.175341 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.175544 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.175756 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.278518 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.278877 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.278962 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.279044 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.279106 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.309132 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:00 crc kubenswrapper[4575]: E1004 04:35:00.309604 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.381987 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.382025 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.382038 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.382054 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.382065 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.484061 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.484098 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.484107 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.484122 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.484132 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.586821 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.586865 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.586876 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.586894 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.586905 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.689148 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.689367 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.689429 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.689498 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.689559 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.791451 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.791479 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.791486 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.791498 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.791507 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.893540 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.893599 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.893612 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.893625 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.893639 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.996103 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.996150 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.996166 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.996186 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:00 crc kubenswrapper[4575]: I1004 04:35:00.996201 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:00Z","lastTransitionTime":"2025-10-04T04:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.046744 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.046886 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:33.046859646 +0000 UTC m=+84.375418460 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.098357 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.098394 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.098405 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.098418 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.098428 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:01Z","lastTransitionTime":"2025-10-04T04:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.148438 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.148487 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.148576 4575 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.148622 4575 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.148665 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:33.148649098 +0000 UTC m=+84.477207912 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.148681 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:33.148673969 +0000 UTC m=+84.477232783 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.200166 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.200208 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.200219 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.200356 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.200378 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:01Z","lastTransitionTime":"2025-10-04T04:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.249832 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.249932 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.250089 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.250136 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.250148 4575 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.250174 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.250208 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.250232 4575 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.250215 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:33.250196183 +0000 UTC m=+84.578754997 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.250385 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:33.250338247 +0000 UTC m=+84.578897061 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.303474 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.303555 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.303574 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.303648 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.303672 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:01Z","lastTransitionTime":"2025-10-04T04:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.309958 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.309996 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.310050 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.310195 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.310302 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:01 crc kubenswrapper[4575]: E1004 04:35:01.310407 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.406403 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.406453 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.406469 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.406491 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.406508 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:01Z","lastTransitionTime":"2025-10-04T04:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.508620 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.508648 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.508658 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.508678 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.508693 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:01Z","lastTransitionTime":"2025-10-04T04:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.610884 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.610915 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.610924 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.610939 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.610948 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:01Z","lastTransitionTime":"2025-10-04T04:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.713448 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.713501 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.713512 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.713530 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.713541 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:01Z","lastTransitionTime":"2025-10-04T04:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.815645 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.815709 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.815720 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.815735 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.815765 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:01Z","lastTransitionTime":"2025-10-04T04:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.918190 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.918235 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.918247 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.918262 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:01 crc kubenswrapper[4575]: I1004 04:35:01.918275 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:01Z","lastTransitionTime":"2025-10-04T04:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.020760 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.021077 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.021229 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.021324 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.021397 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.124483 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.124524 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.124532 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.124545 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.124553 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.227533 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.227604 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.227620 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.227640 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.227652 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.309503 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:02 crc kubenswrapper[4575]: E1004 04:35:02.309640 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.329562 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.329612 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.329622 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.329636 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.329647 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.432795 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.432834 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.432842 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.432858 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.432867 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.535278 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.535864 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.536094 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.536257 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.536401 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.638726 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.638955 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.639021 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.639108 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.639185 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.741405 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.741453 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.741465 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.741482 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.741493 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.844580 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.844669 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.844687 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.844710 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.844729 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.947038 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.947071 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.947080 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.947093 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:02 crc kubenswrapper[4575]: I1004 04:35:02.947101 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:02Z","lastTransitionTime":"2025-10-04T04:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.048997 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.049029 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.049037 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.049049 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.049058 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.151020 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.151068 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.151081 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.151099 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.151110 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.253192 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.253253 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.253289 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.253306 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.253319 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.309277 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.309372 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.309284 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:03 crc kubenswrapper[4575]: E1004 04:35:03.309430 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:03 crc kubenswrapper[4575]: E1004 04:35:03.309692 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:03 crc kubenswrapper[4575]: E1004 04:35:03.309770 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.355760 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.355795 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.355804 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.355817 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.355826 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.457925 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.457961 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.457971 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.457990 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.458001 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.559847 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.559899 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.559910 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.559924 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.559935 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.662275 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.662536 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.662727 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.662909 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.663094 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.766221 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.766283 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.766294 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.766319 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.766332 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.868641 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.868681 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.868692 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.868709 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.868738 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.972054 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.972120 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.972136 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.972156 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:03 crc kubenswrapper[4575]: I1004 04:35:03.972172 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:03Z","lastTransitionTime":"2025-10-04T04:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.074633 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.074672 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.074682 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.074698 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.074709 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.178035 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.178080 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.178090 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.178106 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.178117 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.279827 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.279857 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.279866 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.279879 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.279887 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.309231 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:04 crc kubenswrapper[4575]: E1004 04:35:04.309533 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.382535 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.382567 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.382575 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.382604 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.382615 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.485322 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.485388 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.485405 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.485429 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.485448 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.587322 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.587363 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.587371 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.587385 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.587393 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.689886 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.689938 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.689950 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.689968 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.689980 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.793420 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.793490 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.793503 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.793526 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.793540 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.895799 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.895832 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.895842 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.895856 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.895866 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.997891 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.998650 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.998728 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.998848 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:04 crc kubenswrapper[4575]: I1004 04:35:04.998926 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:04Z","lastTransitionTime":"2025-10-04T04:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.101211 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.101247 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.101257 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.101271 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.101280 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.203363 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.203406 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.203418 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.203434 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.203446 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.305906 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.305937 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.305945 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.305960 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.305971 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.309120 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.309179 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.309142 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:05 crc kubenswrapper[4575]: E1004 04:35:05.309241 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:05 crc kubenswrapper[4575]: E1004 04:35:05.309358 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:05 crc kubenswrapper[4575]: E1004 04:35:05.309693 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.309950 4575 scope.go:117] "RemoveContainer" containerID="0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.408726 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.409049 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.409061 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.409077 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.409089 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.511210 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.511273 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.511288 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.511310 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.511323 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.613856 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.613904 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.613913 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.613927 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.613937 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.716251 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.716315 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.716327 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.716344 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.716358 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.819017 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.819077 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.819090 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.819106 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.819137 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.875064 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.875107 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.875118 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.875136 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.875147 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: E1004 04:35:05.889114 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.894339 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.894380 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.894391 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.894409 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.894422 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: E1004 04:35:05.924046 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.929949 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.929984 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.929996 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.930034 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.930045 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:05 crc kubenswrapper[4575]: E1004 04:35:05.965102 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.985356 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.985396 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.985408 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.985423 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:05 crc kubenswrapper[4575]: I1004 04:35:05.985435 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:05Z","lastTransitionTime":"2025-10-04T04:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: E1004 04:35:06.018767 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.022349 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.022381 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.022400 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.022415 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.022425 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: E1004 04:35:06.034903 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: E1004 04:35:06.035113 4575 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.036963 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.036999 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.037008 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.037023 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.037034 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.139076 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.139116 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.139126 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.139142 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.139154 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.241468 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.241515 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.241528 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.241545 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.241556 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.309499 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:06 crc kubenswrapper[4575]: E1004 04:35:06.309658 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.343950 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.343985 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.343995 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.344010 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.344020 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.446410 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.446438 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.446447 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.446504 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.446515 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.549730 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.549774 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.549782 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.549799 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.549808 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.628224 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/2.log" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.628800 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/1.log" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.631675 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80" exitCode=1 Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.631719 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.631759 4575 scope.go:117] "RemoveContainer" containerID="0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.632486 4575 scope.go:117] "RemoveContainer" containerID="3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80" Oct 04 04:35:06 crc kubenswrapper[4575]: E1004 04:35:06.632685 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.645058 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.652215 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.652523 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.652648 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.652745 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.652844 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.658042 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.676508 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:48Z\\\",\\\"message\\\":\\\"ere:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:34:48.863920 6032 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:34:48.863926 6032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"ices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 04:35:06.350658 6216 services_controller.go:444] Built service openshift-machine-api/cluster-autoscaler-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350666 6216 services_controller.go:445] Built service openshift-machine-api/cluster-autoscaler-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350664 6216 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:35:06.350694 6216 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.688863 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.699407 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.713763 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.724431 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.737865 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.751609 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.755888 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.755917 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.755927 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.755943 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.755958 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.777671 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.792155 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.808007 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.820368 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.833795 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.847412 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.857723 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.857751 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.857760 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.857773 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.857782 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.860376 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.873468 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.959641 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.959673 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.959683 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.959698 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:06 crc kubenswrapper[4575]: I1004 04:35:06.959708 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:06Z","lastTransitionTime":"2025-10-04T04:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.062305 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.062343 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.062353 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.062369 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.062379 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.165000 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.165286 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.165410 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.165545 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.165686 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.268164 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.268207 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.268218 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.268234 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.268245 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.309768 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.309809 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.309823 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:07 crc kubenswrapper[4575]: E1004 04:35:07.309902 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:07 crc kubenswrapper[4575]: E1004 04:35:07.309983 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:07 crc kubenswrapper[4575]: E1004 04:35:07.310065 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.370377 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.370410 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.370420 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.370434 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.370444 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.472485 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.472520 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.472533 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.472552 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.472567 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.574571 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.574620 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.574629 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.574641 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.574649 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.634960 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/2.log" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.678058 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.678122 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.678135 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.678155 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.678169 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.781336 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.781380 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.781392 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.781414 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.781427 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.884493 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.884524 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.884536 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.884551 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.884563 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.986860 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.986918 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.986930 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.986945 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:07 crc kubenswrapper[4575]: I1004 04:35:07.986955 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:07Z","lastTransitionTime":"2025-10-04T04:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.089480 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.089519 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.089531 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.089548 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.089561 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:08Z","lastTransitionTime":"2025-10-04T04:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.192340 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.192463 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.192492 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.192521 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.192543 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:08Z","lastTransitionTime":"2025-10-04T04:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.295131 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.295166 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.295175 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.295189 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.295197 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:08Z","lastTransitionTime":"2025-10-04T04:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.309373 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:08 crc kubenswrapper[4575]: E1004 04:35:08.309543 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.397776 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.397831 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.397844 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.397860 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.397873 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:08Z","lastTransitionTime":"2025-10-04T04:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.501039 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.501077 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.501087 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.501108 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.501122 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:08Z","lastTransitionTime":"2025-10-04T04:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.603465 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.603505 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.603768 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.603787 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.603797 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:08Z","lastTransitionTime":"2025-10-04T04:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.706142 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.706198 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.706207 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.706219 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.706228 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:08Z","lastTransitionTime":"2025-10-04T04:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.808155 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.808228 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.808241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.808256 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.808267 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:08Z","lastTransitionTime":"2025-10-04T04:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.911066 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.911155 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.911176 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.911203 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:08 crc kubenswrapper[4575]: I1004 04:35:08.911223 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:08Z","lastTransitionTime":"2025-10-04T04:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.013018 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.013054 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.013064 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.013079 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.013088 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.115404 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.115443 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.115457 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.115472 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.115482 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.217885 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.217926 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.217935 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.217950 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.217959 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.309342 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:09 crc kubenswrapper[4575]: E1004 04:35:09.309454 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.309653 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:09 crc kubenswrapper[4575]: E1004 04:35:09.309697 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.309786 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:09 crc kubenswrapper[4575]: E1004 04:35:09.309910 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.321859 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.322146 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.322232 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.322318 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.322383 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.323307 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.335497 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.352608 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c2a16b2673f965066b34b4a0da76051613755d8461ced2c8ae075e07d01483a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:48Z\\\",\\\"message\\\":\\\"ere:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:34:48.863920 6032 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:34:48.863926 6032 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:48Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"ices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 04:35:06.350658 6216 services_controller.go:444] Built service openshift-machine-api/cluster-autoscaler-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350666 6216 services_controller.go:445] Built service openshift-machine-api/cluster-autoscaler-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350664 6216 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:35:06.350694 6216 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.366060 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.378771 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.391439 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.402597 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.414730 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.424374 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.424407 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.424417 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.424429 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.424440 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.426070 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.436043 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.444493 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.459576 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.470207 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.479635 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.492402 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.502934 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.512197 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:09Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.526822 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.526859 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.526870 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.526884 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.526893 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.628615 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.628655 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.628667 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.628682 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.628692 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.730986 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.731025 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.731034 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.731047 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.731058 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.833777 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.833841 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.833851 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.833864 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.833927 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.935964 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.936004 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.936015 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.936031 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:09 crc kubenswrapper[4575]: I1004 04:35:09.936044 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:09Z","lastTransitionTime":"2025-10-04T04:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.038384 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.038417 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.038427 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.038444 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.038455 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.141576 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.141630 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.141638 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.141650 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.141658 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.244810 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.244858 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.244870 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.244887 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.244898 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.309576 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:10 crc kubenswrapper[4575]: E1004 04:35:10.309718 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.347216 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.347246 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.347254 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.347267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.347275 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.449338 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.449372 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.449383 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.449399 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.449410 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.551901 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.551960 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.551973 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.551990 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.551999 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.653876 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.653912 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.653922 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.653935 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.653945 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.756035 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.756083 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.756101 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.756121 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.756135 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.858342 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.858377 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.858385 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.858398 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.858409 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.960831 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.960869 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.960879 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.960893 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:10 crc kubenswrapper[4575]: I1004 04:35:10.960904 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:10Z","lastTransitionTime":"2025-10-04T04:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.063032 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.063079 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.063093 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.063108 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.063117 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.165277 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.165315 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.165327 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.165340 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.165348 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.267990 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.268312 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.268323 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.268338 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.268351 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.309778 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.309844 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:11 crc kubenswrapper[4575]: E1004 04:35:11.309911 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.309956 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:11 crc kubenswrapper[4575]: E1004 04:35:11.310079 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:11 crc kubenswrapper[4575]: E1004 04:35:11.310179 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.370566 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.370630 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.370651 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.370669 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.370680 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.472750 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.472788 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.472800 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.472816 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.472827 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.574905 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.574947 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.574959 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.574975 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.574987 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.677160 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.677198 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.677207 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.677222 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.677233 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.779657 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.779687 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.779695 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.779709 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.779718 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.882049 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.882102 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.882112 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.882129 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.882140 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.984218 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.984272 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.984280 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.984292 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:11 crc kubenswrapper[4575]: I1004 04:35:11.984300 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:11Z","lastTransitionTime":"2025-10-04T04:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.086385 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.086667 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.086769 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.086864 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.086955 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:12Z","lastTransitionTime":"2025-10-04T04:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.189252 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.189278 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.189287 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.189298 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.189306 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:12Z","lastTransitionTime":"2025-10-04T04:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.291577 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.291862 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.291979 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.292102 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.292186 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:12Z","lastTransitionTime":"2025-10-04T04:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.309851 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:12 crc kubenswrapper[4575]: E1004 04:35:12.310006 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.394533 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.394569 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.394580 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.394623 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.394634 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:12Z","lastTransitionTime":"2025-10-04T04:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.497315 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.497352 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.497365 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.497380 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.497391 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:12Z","lastTransitionTime":"2025-10-04T04:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.599506 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.599547 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.599561 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.599578 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.599605 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:12Z","lastTransitionTime":"2025-10-04T04:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.701558 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.701642 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.701657 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.701676 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.701713 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:12Z","lastTransitionTime":"2025-10-04T04:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.804075 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.804115 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.804127 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.804143 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.804154 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:12Z","lastTransitionTime":"2025-10-04T04:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.906180 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.906214 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.906225 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.906241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:12 crc kubenswrapper[4575]: I1004 04:35:12.906257 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:12Z","lastTransitionTime":"2025-10-04T04:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.008494 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.008618 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.008633 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.008648 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.008658 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.110822 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.110860 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.110872 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.110888 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.110900 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.213402 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.213448 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.213459 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.213477 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.213490 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.309463 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.309528 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:13 crc kubenswrapper[4575]: E1004 04:35:13.309639 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.310020 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:13 crc kubenswrapper[4575]: E1004 04:35:13.310091 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:13 crc kubenswrapper[4575]: E1004 04:35:13.310144 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.315082 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.315120 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.315133 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.315149 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.315190 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.417871 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.417903 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.417912 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.417926 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.417935 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.521027 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.521073 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.521084 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.521098 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.521107 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.623195 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.623231 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.623239 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.623254 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.623265 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.726007 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.726056 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.726068 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.726082 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.726106 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.827958 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.827999 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.828011 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.828027 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.828038 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.931006 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.931039 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.931049 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.931065 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:13 crc kubenswrapper[4575]: I1004 04:35:13.931076 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:13Z","lastTransitionTime":"2025-10-04T04:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.033371 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.033400 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.033410 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.033423 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.033432 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.136219 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.136260 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.136271 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.136290 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.136302 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.238340 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.238379 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.238389 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.238405 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.238416 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.309653 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:14 crc kubenswrapper[4575]: E1004 04:35:14.309794 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.340539 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.340601 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.340614 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.340631 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.340644 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.442389 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.442441 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.442453 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.442471 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.442482 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.544790 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.544827 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.544838 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.544853 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.544864 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.647086 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.647137 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.647173 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.647189 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.647201 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.750038 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.750608 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.750781 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.751055 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.751158 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.853394 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.853643 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.853722 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.853790 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.853899 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.956157 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.956184 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.956193 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.956205 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:14 crc kubenswrapper[4575]: I1004 04:35:14.956214 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:14Z","lastTransitionTime":"2025-10-04T04:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.058082 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.058295 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.058377 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.058457 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.058532 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.160225 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.160257 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.160267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.160281 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.160291 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.262682 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.262722 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.262730 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.262746 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.262755 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.309524 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:15 crc kubenswrapper[4575]: E1004 04:35:15.309680 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.309739 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:15 crc kubenswrapper[4575]: E1004 04:35:15.309822 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.309535 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:15 crc kubenswrapper[4575]: E1004 04:35:15.309892 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.364667 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.364723 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.364733 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.364745 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.364752 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.466771 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.466804 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.466814 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.466829 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.466840 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.569424 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.569484 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.569495 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.569510 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.569521 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.671132 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.671171 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.671178 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.671193 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.671201 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.773241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.773628 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.773705 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.773773 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.773836 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.786628 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:15 crc kubenswrapper[4575]: E1004 04:35:15.786763 4575 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:35:15 crc kubenswrapper[4575]: E1004 04:35:15.786837 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs podName:026aaa63-845a-45bb-bc23-7b7c3024eead nodeName:}" failed. No retries permitted until 2025-10-04 04:35:47.786816681 +0000 UTC m=+99.115375495 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs") pod "network-metrics-daemon-9nvms" (UID: "026aaa63-845a-45bb-bc23-7b7c3024eead") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.875923 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.875952 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.875963 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.875978 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.875990 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.978298 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.978332 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.978343 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.978358 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:15 crc kubenswrapper[4575]: I1004 04:35:15.978369 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:15Z","lastTransitionTime":"2025-10-04T04:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.070833 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.070861 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.070869 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.070881 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.070889 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: E1004 04:35:16.085081 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.088007 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.088035 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.088045 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.088059 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.088068 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: E1004 04:35:16.098496 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.101454 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.101500 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.101510 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.101523 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.101531 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: E1004 04:35:16.112052 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.114950 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.114971 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.114979 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.114992 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.115001 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: E1004 04:35:16.125195 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.127883 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.127911 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.127922 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.127937 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.127947 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: E1004 04:35:16.138030 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: E1004 04:35:16.138147 4575 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.139322 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.139345 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.139353 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.139367 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.139375 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.241648 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.241882 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.241955 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.242014 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.242079 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.309498 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:16 crc kubenswrapper[4575]: E1004 04:35:16.309680 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.344231 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.344309 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.344321 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.344338 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.344349 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.446398 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.446666 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.446752 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.446831 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.446907 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.549366 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.549623 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.549709 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.549778 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.549835 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.651870 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.652109 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.652175 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.652241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.652311 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.754477 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.754608 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.754619 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.754631 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.754640 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.856455 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.856712 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.856802 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.856895 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.856966 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.878857 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.879757 4575 scope.go:117] "RemoveContainer" containerID="3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80" Oct 04 04:35:16 crc kubenswrapper[4575]: E1004 04:35:16.880060 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.893506 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.905330 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.914329 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.927438 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.941057 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.959453 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.959490 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.959502 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.959519 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.959530 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:16Z","lastTransitionTime":"2025-10-04T04:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.962827 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"ices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 04:35:06.350658 6216 services_controller.go:444] Built service openshift-machine-api/cluster-autoscaler-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350666 6216 services_controller.go:445] Built service openshift-machine-api/cluster-autoscaler-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350664 6216 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:35:06.350694 6216 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.973161 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.984211 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:16 crc kubenswrapper[4575]: I1004 04:35:16.995523 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.006190 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.017254 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.028296 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.037250 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.053860 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.061505 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.061552 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.061561 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.061579 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.061604 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.065008 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.075459 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.086451 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.163529 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.163561 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.163569 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.163601 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.163611 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.265954 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.266018 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.266061 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.266077 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.266089 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.310765 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:17 crc kubenswrapper[4575]: E1004 04:35:17.310889 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.311239 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:17 crc kubenswrapper[4575]: E1004 04:35:17.311304 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.311365 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:17 crc kubenswrapper[4575]: E1004 04:35:17.311425 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.368185 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.368215 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.368223 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.368237 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.368247 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.470121 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.470158 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.470170 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.470199 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.470213 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.571833 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.571868 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.571878 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.571891 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.571901 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.674155 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.674186 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.674196 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.674209 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.674218 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.776891 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.776928 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.776938 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.776955 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.776968 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.878678 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.878716 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.878724 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.878738 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.878749 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.980534 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.980562 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.980570 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.980597 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:17 crc kubenswrapper[4575]: I1004 04:35:17.980610 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:17Z","lastTransitionTime":"2025-10-04T04:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.083031 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.083082 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.083094 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.083114 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.083127 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:18Z","lastTransitionTime":"2025-10-04T04:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.186139 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.186168 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.186176 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.186189 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.186198 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:18Z","lastTransitionTime":"2025-10-04T04:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.289032 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.289319 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.289400 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.289493 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.289574 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:18Z","lastTransitionTime":"2025-10-04T04:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.308851 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:18 crc kubenswrapper[4575]: E1004 04:35:18.309041 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.392179 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.392215 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.392226 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.392240 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.392251 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:18Z","lastTransitionTime":"2025-10-04T04:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.494672 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.494711 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.494723 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.494737 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.494750 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:18Z","lastTransitionTime":"2025-10-04T04:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.596393 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.596429 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.596438 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.596452 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.596462 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:18Z","lastTransitionTime":"2025-10-04T04:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.698148 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.698174 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.698183 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.698194 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.698203 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:18Z","lastTransitionTime":"2025-10-04T04:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.800027 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.800281 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.800411 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.800507 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.800580 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:18Z","lastTransitionTime":"2025-10-04T04:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.903257 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.903303 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.903315 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.903336 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:18 crc kubenswrapper[4575]: I1004 04:35:18.903349 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:18Z","lastTransitionTime":"2025-10-04T04:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.005917 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.005964 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.005976 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.005995 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.006007 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.108615 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.108661 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.108672 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.108692 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.108703 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.211059 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.211096 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.211105 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.211118 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.211127 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.309368 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:19 crc kubenswrapper[4575]: E1004 04:35:19.309506 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.309373 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:19 crc kubenswrapper[4575]: E1004 04:35:19.309674 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.309872 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:19 crc kubenswrapper[4575]: E1004 04:35:19.309928 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.314702 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.314747 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.314758 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.314870 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.314885 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.324925 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.335958 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.345909 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.357861 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.369509 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.387452 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"ices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 04:35:06.350658 6216 services_controller.go:444] Built service openshift-machine-api/cluster-autoscaler-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350666 6216 services_controller.go:445] Built service openshift-machine-api/cluster-autoscaler-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350664 6216 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:35:06.350694 6216 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.401818 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.413847 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.418224 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.418341 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.418353 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.418365 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.418375 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.425065 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.434655 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.445973 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.457498 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.472763 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.483005 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.497923 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.514380 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.521091 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.521126 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.521135 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.521149 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.521158 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.528223 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.623610 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.623657 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.623669 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.623689 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.623704 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.725934 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.726022 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.726034 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.726050 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.726061 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.828407 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.829078 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.829113 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.829135 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.829147 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.931435 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.931742 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.931852 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.931951 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:19 crc kubenswrapper[4575]: I1004 04:35:19.932029 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:19Z","lastTransitionTime":"2025-10-04T04:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.034130 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.034185 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.034195 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.034210 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.034219 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.136659 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.136909 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.136975 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.137041 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.137112 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.239298 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.239338 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.239346 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.239360 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.239368 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.309163 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:20 crc kubenswrapper[4575]: E1004 04:35:20.309285 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.341797 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.341845 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.341855 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.341867 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.341876 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.443911 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.443946 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.443954 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.443967 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.443975 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.550528 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.550656 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.550680 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.550698 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.550710 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.652689 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.652720 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.652730 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.652744 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.652755 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.754558 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.754618 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.754630 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.754647 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.754658 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.857726 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.857760 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.857769 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.857784 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.857797 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.960154 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.960210 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.960219 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.960235 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:20 crc kubenswrapper[4575]: I1004 04:35:20.960245 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:20Z","lastTransitionTime":"2025-10-04T04:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.062836 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.062923 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.062943 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.062966 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.062979 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.165223 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.165258 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.165267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.165283 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.165294 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.267522 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.267570 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.267611 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.267625 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.267634 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.309251 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.309325 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:21 crc kubenswrapper[4575]: E1004 04:35:21.309378 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:21 crc kubenswrapper[4575]: E1004 04:35:21.309469 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.309272 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:21 crc kubenswrapper[4575]: E1004 04:35:21.309556 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.369562 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.369609 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.369620 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.369636 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.369647 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.472402 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.472458 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.472470 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.472485 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.472495 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.574548 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.574598 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.574611 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.574627 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.574665 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.676689 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.676730 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.676749 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.676765 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.676777 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.778741 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.778777 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.778789 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.778805 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.778816 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.881300 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.881345 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.881366 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.881390 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.881399 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.983495 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.983539 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.983547 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.983560 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:21 crc kubenswrapper[4575]: I1004 04:35:21.983569 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:21Z","lastTransitionTime":"2025-10-04T04:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.085659 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.085684 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.085692 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.085704 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.085712 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:22Z","lastTransitionTime":"2025-10-04T04:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.188193 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.188256 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.188273 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.188299 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.188316 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:22Z","lastTransitionTime":"2025-10-04T04:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.291058 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.291098 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.291109 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.291124 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.291136 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:22Z","lastTransitionTime":"2025-10-04T04:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.309167 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:22 crc kubenswrapper[4575]: E1004 04:35:22.309329 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.394616 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.394649 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.394658 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.394670 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.394679 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:22Z","lastTransitionTime":"2025-10-04T04:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.497032 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.497067 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.497077 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.497095 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.497107 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:22Z","lastTransitionTime":"2025-10-04T04:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.598863 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.598899 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.598908 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.598922 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.598932 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:22Z","lastTransitionTime":"2025-10-04T04:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.701612 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.701648 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.701661 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.701676 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.701687 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:22Z","lastTransitionTime":"2025-10-04T04:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.804127 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.804168 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.804180 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.804197 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.804209 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:22Z","lastTransitionTime":"2025-10-04T04:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.906935 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.907027 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.907044 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.907066 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:22 crc kubenswrapper[4575]: I1004 04:35:22.907088 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:22Z","lastTransitionTime":"2025-10-04T04:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.009088 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.009120 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.009128 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.009141 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.009149 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.111082 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.111112 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.111122 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.111134 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.111143 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.212748 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.212798 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.212810 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.212826 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.212837 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.309212 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.309268 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:23 crc kubenswrapper[4575]: E1004 04:35:23.309390 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:23 crc kubenswrapper[4575]: E1004 04:35:23.309503 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.309163 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:23 crc kubenswrapper[4575]: E1004 04:35:23.309724 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.314749 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.314791 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.314800 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.314811 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.314820 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.417219 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.417261 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.417270 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.417285 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.417294 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.519496 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.519547 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.519563 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.519603 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.519618 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.622205 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.622248 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.622258 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.622273 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.622284 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.682339 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sj5hq_31197ccb-a3f9-4130-bca4-ca1c3d3ee065/kube-multus/0.log" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.682397 4575 generic.go:334] "Generic (PLEG): container finished" podID="31197ccb-a3f9-4130-bca4-ca1c3d3ee065" containerID="adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649" exitCode=1 Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.682432 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sj5hq" event={"ID":"31197ccb-a3f9-4130-bca4-ca1c3d3ee065","Type":"ContainerDied","Data":"adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.682861 4575 scope.go:117] "RemoveContainer" containerID="adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.693392 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.702282 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.718048 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.724890 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.724926 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.724936 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.724949 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.724960 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.730055 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.748063 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"ices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 04:35:06.350658 6216 services_controller.go:444] Built service openshift-machine-api/cluster-autoscaler-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350666 6216 services_controller.go:445] Built service openshift-machine-api/cluster-autoscaler-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350664 6216 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:35:06.350694 6216 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.764158 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.776126 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.786101 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.796279 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:22Z\\\",\\\"message\\\":\\\"2025-10-04T04:34:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a\\\\n2025-10-04T04:34:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a to /host/opt/cni/bin/\\\\n2025-10-04T04:34:37Z [verbose] multus-daemon started\\\\n2025-10-04T04:34:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:35:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.806053 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.815536 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.824720 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.827941 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.827970 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.827977 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.827991 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.828000 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.836546 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.845249 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.854165 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.864642 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.875719 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.930728 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.930755 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.930763 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.930791 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:23 crc kubenswrapper[4575]: I1004 04:35:23.930800 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:23Z","lastTransitionTime":"2025-10-04T04:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.033555 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.033635 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.033649 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.033663 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.033674 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.135751 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.135806 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.135817 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.135835 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.135847 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.238675 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.238719 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.238733 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.238748 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.238758 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.309167 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:24 crc kubenswrapper[4575]: E1004 04:35:24.309451 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.340893 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.340940 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.340952 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.340965 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.340974 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.443154 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.443214 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.443226 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.443240 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.443254 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.546737 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.546804 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.546827 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.546855 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.546877 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.649341 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.649422 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.649450 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.649479 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.649501 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.688234 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sj5hq_31197ccb-a3f9-4130-bca4-ca1c3d3ee065/kube-multus/0.log" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.688292 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sj5hq" event={"ID":"31197ccb-a3f9-4130-bca4-ca1c3d3ee065","Type":"ContainerStarted","Data":"ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.712163 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.731875 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.745116 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.752060 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.752106 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.752119 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.752135 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.752148 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.763019 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.774617 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.784525 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.798182 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.808871 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.816989 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.829858 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.843536 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.855174 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.855220 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.855232 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.855246 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.855255 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.865210 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"ices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 04:35:06.350658 6216 services_controller.go:444] Built service openshift-machine-api/cluster-autoscaler-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350666 6216 services_controller.go:445] Built service openshift-machine-api/cluster-autoscaler-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350664 6216 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:35:06.350694 6216 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.878737 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.889993 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.901767 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.912905 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.924182 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:22Z\\\",\\\"message\\\":\\\"2025-10-04T04:34:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a\\\\n2025-10-04T04:34:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a to /host/opt/cni/bin/\\\\n2025-10-04T04:34:37Z [verbose] multus-daemon started\\\\n2025-10-04T04:34:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:35:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.956795 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.956830 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.956839 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.956855 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:24 crc kubenswrapper[4575]: I1004 04:35:24.956864 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:24Z","lastTransitionTime":"2025-10-04T04:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.058721 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.058785 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.058796 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.058810 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.058821 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.161313 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.161353 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.161363 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.161379 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.161390 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.263608 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.263670 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.263686 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.263706 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.263724 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.309321 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.309355 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.309400 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:25 crc kubenswrapper[4575]: E1004 04:35:25.309575 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:25 crc kubenswrapper[4575]: E1004 04:35:25.309712 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:25 crc kubenswrapper[4575]: E1004 04:35:25.309841 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.366168 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.366213 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.366225 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.366243 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.366255 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.468515 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.468569 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.468580 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.468614 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.468623 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.571313 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.571341 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.571349 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.571361 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.571368 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.674442 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.674508 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.674530 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.674561 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.674615 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.776543 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.776580 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.776611 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.776626 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.776637 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.879212 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.879256 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.879268 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.879285 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.879298 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.982043 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.982080 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.982089 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.982101 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:25 crc kubenswrapper[4575]: I1004 04:35:25.982112 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:25Z","lastTransitionTime":"2025-10-04T04:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.084817 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.084860 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.084874 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.084892 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.084906 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.187166 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.187193 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.187201 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.187213 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.187221 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.229301 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.229360 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.229376 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.229399 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.229415 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: E1004 04:35:26.243231 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.246314 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.246346 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.246356 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.246377 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.246389 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: E1004 04:35:26.259060 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.262647 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.262708 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.262720 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.262736 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.262783 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: E1004 04:35:26.273385 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.276526 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.276580 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.276600 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.276612 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.276621 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: E1004 04:35:26.288254 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.291242 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.291273 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.291281 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.291298 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.291309 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: E1004 04:35:26.303020 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:26 crc kubenswrapper[4575]: E1004 04:35:26.303142 4575 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.304535 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.304578 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.304602 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.304618 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.304627 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.309748 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:26 crc kubenswrapper[4575]: E1004 04:35:26.309866 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.406832 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.406869 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.406877 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.406893 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.406902 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.509510 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.509552 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.509561 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.509576 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.509616 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.612300 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.612342 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.612353 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.612369 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.612382 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.714546 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.714609 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.714622 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.714641 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.714655 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.817206 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.817247 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.817263 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.817279 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.817288 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.919859 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.920106 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.920267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.920368 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:26 crc kubenswrapper[4575]: I1004 04:35:26.920446 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:26Z","lastTransitionTime":"2025-10-04T04:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.022252 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.022305 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.022319 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.022337 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.022349 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.124896 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.124955 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.124967 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.124983 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.124993 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.226988 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.227247 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.227480 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.227839 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.228033 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.309239 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.309360 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.309449 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:27 crc kubenswrapper[4575]: E1004 04:35:27.309542 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:27 crc kubenswrapper[4575]: E1004 04:35:27.309757 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:27 crc kubenswrapper[4575]: E1004 04:35:27.310191 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.310441 4575 scope.go:117] "RemoveContainer" containerID="3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.332570 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.332617 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.332628 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.332643 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.332652 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.439358 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.439433 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.439444 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.439458 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.439471 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.542375 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.542420 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.542436 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.542457 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.542474 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.645016 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.645064 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.645081 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.645105 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.645122 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.699842 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/2.log" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.702761 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.747267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.747349 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.747368 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.747387 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.747402 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.849665 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.849709 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.849721 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.849741 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.849754 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.952131 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.952172 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.952186 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.952201 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:27 crc kubenswrapper[4575]: I1004 04:35:27.952213 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:27Z","lastTransitionTime":"2025-10-04T04:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.053688 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.053715 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.053723 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.053735 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.053743 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.156418 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.156446 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.156457 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.156472 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.156482 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.258749 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.258819 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.258840 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.258874 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.258893 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.309181 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:28 crc kubenswrapper[4575]: E1004 04:35:28.309302 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.360501 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.360540 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.360553 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.360570 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.360598 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.462747 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.462791 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.462801 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.462822 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.462837 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.565173 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.565219 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.565231 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.565249 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.565261 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.668043 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.668075 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.668086 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.668102 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.668114 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.707905 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/3.log" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.708619 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/2.log" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.711152 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" exitCode=1 Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.711189 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.711222 4575 scope.go:117] "RemoveContainer" containerID="3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.712055 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:35:28 crc kubenswrapper[4575]: E1004 04:35:28.712280 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.729303 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.746147 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.757280 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.771101 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.771136 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.771151 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.771165 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.771173 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.775696 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.786491 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.795063 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.809848 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.820352 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.832142 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.844568 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.856177 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.872139 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"ices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 04:35:06.350658 6216 services_controller.go:444] Built service openshift-machine-api/cluster-autoscaler-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350666 6216 services_controller.go:445] Built service openshift-machine-api/cluster-autoscaler-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350664 6216 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:35:06.350694 6216 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:28Z\\\",\\\"message\\\":\\\"vices.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI1004 04:35:28.235805 6526 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 04:35:28.235861 6526 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.873306 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.873335 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.873344 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.873358 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.873367 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.884127 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.894734 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.905369 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.917382 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.938899 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:22Z\\\",\\\"message\\\":\\\"2025-10-04T04:34:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a\\\\n2025-10-04T04:34:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a to /host/opt/cni/bin/\\\\n2025-10-04T04:34:37Z [verbose] multus-daemon started\\\\n2025-10-04T04:34:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:35:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.975477 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.975517 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.975528 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.975553 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:28 crc kubenswrapper[4575]: I1004 04:35:28.975565 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:28Z","lastTransitionTime":"2025-10-04T04:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.078090 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.078161 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.078176 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.078192 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.078202 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.180483 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.180513 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.180521 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.180533 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.180542 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.282909 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.282946 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.282957 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.282971 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.282981 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.309467 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:29 crc kubenswrapper[4575]: E1004 04:35:29.309625 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.309640 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.309657 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:29 crc kubenswrapper[4575]: E1004 04:35:29.309717 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:29 crc kubenswrapper[4575]: E1004 04:35:29.309774 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.320109 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.332097 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.341957 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.355031 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.365344 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.375876 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.384752 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.384781 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.384790 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.384804 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.384815 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.390371 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.401513 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.410964 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.422489 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.432885 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.449234 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3062017193a2ec030f78bdb834a5f3280da833247b7de4866d9d8ab36122da80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:06Z\\\",\\\"message\\\":\\\"ices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 04:35:06.350658 6216 services_controller.go:444] Built service openshift-machine-api/cluster-autoscaler-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350666 6216 services_controller.go:445] Built service openshift-machine-api/cluster-autoscaler-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1004 04:35:06.350664 6216 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:35:06.350694 6216 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:28Z\\\",\\\"message\\\":\\\"vices.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI1004 04:35:28.235805 6526 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 04:35:28.235861 6526 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.461536 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.472078 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.485361 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.486801 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.486835 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.486875 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.486895 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.486950 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.495730 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.506141 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:22Z\\\",\\\"message\\\":\\\"2025-10-04T04:34:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a\\\\n2025-10-04T04:34:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a to /host/opt/cni/bin/\\\\n2025-10-04T04:34:37Z [verbose] multus-daemon started\\\\n2025-10-04T04:34:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:35:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.588904 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.588942 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.588952 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.588968 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.588978 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.691303 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.691853 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.691954 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.692017 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.692075 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.716146 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/3.log" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.719901 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:35:29 crc kubenswrapper[4575]: E1004 04:35:29.720242 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.731957 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.740860 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.753150 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.763426 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.782268 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:28Z\\\",\\\"message\\\":\\\"vices.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI1004 04:35:28.235805 6526 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 04:35:28.235861 6526 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.794164 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.794206 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.794221 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.794241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.794256 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.796448 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.809341 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.823700 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.838845 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:22Z\\\",\\\"message\\\":\\\"2025-10-04T04:34:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a\\\\n2025-10-04T04:34:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a to /host/opt/cni/bin/\\\\n2025-10-04T04:34:37Z [verbose] multus-daemon started\\\\n2025-10-04T04:34:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:35:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.852016 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.862496 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.871924 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.884660 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.893903 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.896248 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.896272 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.896281 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.896293 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.896301 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.904651 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.914996 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.925218 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.998776 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.998805 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.998954 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.998972 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:29 crc kubenswrapper[4575]: I1004 04:35:29.998981 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:29Z","lastTransitionTime":"2025-10-04T04:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.100964 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.101019 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.101029 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.101042 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.101051 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:30Z","lastTransitionTime":"2025-10-04T04:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.203642 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.203705 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.203718 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.203734 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.203745 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:30Z","lastTransitionTime":"2025-10-04T04:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.305952 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.306177 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.306267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.306399 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.306487 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:30Z","lastTransitionTime":"2025-10-04T04:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.309140 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:30 crc kubenswrapper[4575]: E1004 04:35:30.309249 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.409230 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.409506 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.409586 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.409700 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.409794 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:30Z","lastTransitionTime":"2025-10-04T04:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.512666 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.512709 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.512720 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.512737 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.512750 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:30Z","lastTransitionTime":"2025-10-04T04:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.614651 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.614694 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.614707 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.614722 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.614731 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:30Z","lastTransitionTime":"2025-10-04T04:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.717524 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.717569 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.717579 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.717621 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.717634 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:30Z","lastTransitionTime":"2025-10-04T04:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.820584 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.820633 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.820642 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.820654 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.820664 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:30Z","lastTransitionTime":"2025-10-04T04:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.923563 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.923632 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.923649 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.923671 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:30 crc kubenswrapper[4575]: I1004 04:35:30.923686 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:30Z","lastTransitionTime":"2025-10-04T04:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.027154 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.027200 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.027210 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.027226 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.027237 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.130477 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.130560 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.130583 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.130658 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.130673 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.233018 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.233072 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.233079 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.233091 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.233100 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.308990 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:31 crc kubenswrapper[4575]: E1004 04:35:31.309111 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.309268 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:31 crc kubenswrapper[4575]: E1004 04:35:31.309312 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.309419 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:31 crc kubenswrapper[4575]: E1004 04:35:31.309487 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.320046 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.335804 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.335842 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.335851 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.335865 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.335874 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.438140 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.438178 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.438189 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.438203 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.438213 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.541040 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.541068 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.541077 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.541089 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.541098 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.643272 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.643566 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.643709 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.643816 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.643909 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.747306 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.747367 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.747387 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.747406 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.747422 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.854480 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.854780 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.855313 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.855460 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.855631 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.958014 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.958338 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.958439 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.958539 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:31 crc kubenswrapper[4575]: I1004 04:35:31.958656 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:31Z","lastTransitionTime":"2025-10-04T04:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.061494 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.061860 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.061952 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.062058 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.062141 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.164678 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.164720 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.164729 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.164744 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.164753 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.267500 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.267553 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.267562 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.267574 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.267607 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.309775 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:32 crc kubenswrapper[4575]: E1004 04:35:32.309936 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.369382 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.369419 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.369428 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.369441 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.369449 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.472409 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.472453 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.472463 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.472477 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.472488 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.574906 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.574947 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.574958 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.574976 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.574987 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.677268 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.677314 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.677324 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.677337 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.677347 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.779525 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.779565 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.779576 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.779619 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.779638 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.893795 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.893855 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.893881 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.893938 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.893953 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.996223 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.996258 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.996267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.996280 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:32 crc kubenswrapper[4575]: I1004 04:35:32.996289 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:32Z","lastTransitionTime":"2025-10-04T04:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.066645 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.066908 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.066890005 +0000 UTC m=+148.395448819 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.098209 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.098241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.098249 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.098261 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.098271 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:33Z","lastTransitionTime":"2025-10-04T04:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.168195 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.168315 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.168361 4575 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.168455 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.168434382 +0000 UTC m=+148.496993256 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.168473 4575 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.168544 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.168522894 +0000 UTC m=+148.497081738 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.200718 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.200754 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.200764 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.200776 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.200786 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:33Z","lastTransitionTime":"2025-10-04T04:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.269373 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.269534 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.269650 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.269663 4575 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.269756 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.269770 4575 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.269778 4575 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.269827 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.269812373 +0000 UTC m=+148.598371187 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.270005 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.269997328 +0000 UTC m=+148.598556142 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.270109 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.303193 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.303254 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.303266 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.303284 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.303294 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:33Z","lastTransitionTime":"2025-10-04T04:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.309818 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.309855 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.309898 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.309952 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.310037 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:33 crc kubenswrapper[4575]: E1004 04:35:33.310156 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.405517 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.405564 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.405575 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.405605 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.405618 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:33Z","lastTransitionTime":"2025-10-04T04:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.508790 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.508863 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.508874 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.508895 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.508909 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:33Z","lastTransitionTime":"2025-10-04T04:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.611741 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.611816 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.611838 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.611907 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.611924 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:33Z","lastTransitionTime":"2025-10-04T04:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.714300 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.714333 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.714341 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.714353 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.714363 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:33Z","lastTransitionTime":"2025-10-04T04:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.817100 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.817138 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.817149 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.817164 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.817175 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:33Z","lastTransitionTime":"2025-10-04T04:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.921261 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.921578 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.921613 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.921637 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:33 crc kubenswrapper[4575]: I1004 04:35:33.921650 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:33Z","lastTransitionTime":"2025-10-04T04:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.023982 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.024025 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.024037 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.024053 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.024066 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.126118 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.126153 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.126161 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.126173 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.126188 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.228662 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.228703 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.228714 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.228727 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.228736 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.309742 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:34 crc kubenswrapper[4575]: E1004 04:35:34.309878 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.325348 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.331227 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.331269 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.331281 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.331302 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.331313 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.434043 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.434108 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.434131 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.434162 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.434184 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.536398 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.536447 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.536485 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.536505 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.536518 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.640049 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.640289 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.640374 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.640480 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.640619 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.743114 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.743346 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.743425 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.743523 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.743622 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.846785 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.846826 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.846837 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.846855 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.846866 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.948808 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.948841 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.948852 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.948887 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:34 crc kubenswrapper[4575]: I1004 04:35:34.948897 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:34Z","lastTransitionTime":"2025-10-04T04:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.051020 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.051065 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.051076 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.051093 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.051104 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.153919 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.153975 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.153988 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.154004 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.154016 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.257242 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.257291 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.257304 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.257323 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.257336 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.309276 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.309312 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.309395 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:35 crc kubenswrapper[4575]: E1004 04:35:35.309423 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:35 crc kubenswrapper[4575]: E1004 04:35:35.309730 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:35 crc kubenswrapper[4575]: E1004 04:35:35.309991 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.360037 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.360084 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.360095 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.360108 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.360117 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.462029 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.462290 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.462355 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.462413 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.462473 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.564432 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.564469 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.564481 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.564496 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.564508 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.667463 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.667507 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.667517 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.667531 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.667542 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.770361 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.770392 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.770407 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.770459 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.770469 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.872804 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.872841 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.872857 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.872870 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.872879 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.975711 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.975747 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.975759 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.975775 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:35 crc kubenswrapper[4575]: I1004 04:35:35.975787 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:35Z","lastTransitionTime":"2025-10-04T04:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.077921 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.077947 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.077956 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.077967 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.077977 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.180342 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.180372 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.180381 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.180395 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.180403 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.282501 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.282533 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.282545 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.282561 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.282572 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.309042 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:36 crc kubenswrapper[4575]: E1004 04:35:36.309165 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.330915 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.330947 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.330956 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.330971 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.330980 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: E1004 04:35:36.342492 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.345765 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.345925 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.346008 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.346120 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.346223 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: E1004 04:35:36.357215 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.361253 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.361285 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.361293 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.361308 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.361316 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: E1004 04:35:36.374140 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.377907 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.377944 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.377953 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.377966 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.377975 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: E1004 04:35:36.390608 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.393499 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.393649 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.393719 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.393795 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.393867 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: E1004 04:35:36.410557 4575 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27d90f2a-0402-4831-8abb-70fda13695ec\\\",\\\"systemUUID\\\":\\\"1317216c-ca93-49ac-8b83-e45b26a19814\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:36Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:36 crc kubenswrapper[4575]: E1004 04:35:36.411004 4575 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.413003 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.413032 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.413039 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.413054 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.413065 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.515367 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.515407 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.515416 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.515430 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.515439 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.617922 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.617966 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.617974 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.617998 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.618008 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.720739 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.720782 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.720812 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.720828 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.720838 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.822567 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.822620 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.822636 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.822654 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.822672 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.924966 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.925011 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.925035 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.925061 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:36 crc kubenswrapper[4575]: I1004 04:35:36.925076 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:36Z","lastTransitionTime":"2025-10-04T04:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.027173 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.027217 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.027229 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.027246 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.027258 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.129102 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.129147 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.129160 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.129173 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.129181 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.231890 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.231939 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.231949 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.231966 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.231977 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.309803 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.309846 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.309825 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:37 crc kubenswrapper[4575]: E1004 04:35:37.309948 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:37 crc kubenswrapper[4575]: E1004 04:35:37.310028 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:37 crc kubenswrapper[4575]: E1004 04:35:37.310153 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.334392 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.334434 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.334449 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.334467 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.334478 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.437069 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.437899 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.437935 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.437952 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.437964 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.540366 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.540415 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.540429 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.540446 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.540458 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.643810 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.643847 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.643858 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.643873 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.643885 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.745733 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.745828 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.745837 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.745849 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.745858 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.848122 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.848167 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.848179 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.848198 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.848210 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.950962 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.951026 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.951038 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.951054 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:37 crc kubenswrapper[4575]: I1004 04:35:37.951068 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:37Z","lastTransitionTime":"2025-10-04T04:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.053322 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.053388 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.053405 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.053423 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.053437 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.155948 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.155988 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.155998 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.156012 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.156021 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.257954 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.257988 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.257997 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.258009 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.258019 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.309469 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:38 crc kubenswrapper[4575]: E1004 04:35:38.309610 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.360469 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.360507 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.360520 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.360536 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.360551 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.463223 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.463278 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.463289 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.463304 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.463336 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.565724 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.565820 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.565846 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.565893 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.565905 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.668714 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.668767 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.668781 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.668799 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.668820 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.770907 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.770955 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.770967 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.770985 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.770997 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.873952 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.873995 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.874014 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.874032 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.874043 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.976967 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.977011 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.977020 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.977035 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:38 crc kubenswrapper[4575]: I1004 04:35:38.977045 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:38Z","lastTransitionTime":"2025-10-04T04:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.083031 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.083090 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.083107 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.083126 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.083147 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:39Z","lastTransitionTime":"2025-10-04T04:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.185883 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.185928 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.185937 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.185951 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.185961 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:39Z","lastTransitionTime":"2025-10-04T04:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.288722 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.288758 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.288766 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.288778 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.288787 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:39Z","lastTransitionTime":"2025-10-04T04:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.308951 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.309038 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:39 crc kubenswrapper[4575]: E1004 04:35:39.309093 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:39 crc kubenswrapper[4575]: E1004 04:35:39.309269 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.309322 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:39 crc kubenswrapper[4575]: E1004 04:35:39.309374 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.326842 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8b12cb5-29cb-4eab-9c0b-72f03b000666\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2324c47409970f6d0d6987416be0be92476c01bce8e9095f77f3db1ab43496e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d397efaf62cb3c1510b3ab396d1efe96755c089882de5e5bb528e7e27406ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94cfb6f9aa2a7971613ce6eb1d6e3dd6769e59178b1c2a3abc3f9813e74a5eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7110dda2fde5e5727d5164bba3095dd2c4fb72a47ea83fd5685a61e3eab09841\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdde181cec7a5f92c6ef7871aa9db54a7601367f04cfc335046c93e3087bc4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56762753ddca856cbb3ffcd278c916cd3eaceb1a98c370d4dbcb2995a4c14551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://56762753ddca856cbb3ffcd278c916cd3eaceb1a98c370d4dbcb2995a4c14551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://763c09570e509b049afbc322da36072f418f80432fc64dee763c74b826805822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://763c09570e509b049afbc322da36072f418f80432fc64dee763c74b826805822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://974491db5db16e05d1724d1649e3ddef7dacfda89412ea9fa8114882a93d87c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974491db5db16e05d1724d1649e3ddef7dacfda89412ea9fa8114882a93d87c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.340445 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6667d8-06ff-4d44-b1eb-1982a71930b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25a9cbfac178fa781d8ad67ffb1c8158c0df6ccf52435b50f1e251c1ae9ed50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31739ecde5433c74a5008826cc142d52145ab72e3ed1421750b7f0f56c7c7c1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c170ea6ff0013f87b488190ea21eb309d46849ebe964bb4f1e850ae52cebb25f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a6523a0b08e09aa4c516e6d54973cfd344c89f468567a99216d23ff70084098\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd2f6a94a1ebcd37e201f01efbfa0dcd46a9a83792649db205227fd082d11c17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:34:23Z\\\",\\\"message\\\":\\\"W1004 04:34:12.573128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 04:34:12.573463 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759552452 cert, and key in /tmp/serving-cert-2158792664/serving-signer.crt, /tmp/serving-cert-2158792664/serving-signer.key\\\\nI1004 04:34:12.891377 1 observer_polling.go:159] Starting file observer\\\\nW1004 04:34:12.894141 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 04:34:12.894274 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 04:34:12.895391 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2158792664/tls.crt::/tmp/serving-cert-2158792664/tls.key\\\\\\\"\\\\nF1004 04:34:23.255992 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3249848c3f60ef9c9a756a9dcd8f5ae7a92937ff6c02be09cb5eada71cc753a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d3ee845e9374d238476ad17981ccb30e29e33aa509607e75811287aefbdf36d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.351196 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0d8bf4b-bd65-486b-9c54-9f792895c27f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8fd0fcaec368fbd657b6b0d7b9774267e04e8c5ab183850c47c78d23b92dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc5b95576619c5c87842777d9a7e9851ca0a50d6e5269420a42565e669c7e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d224bfe28222fdbb4b855172582fa523d13267f26209bf166f4d2d152b2be266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://231a09c51948ee2e2e1434b687501571b478a15f3c2a4d5ed5d57c7b4f7a0772\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.360097 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2q7bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cf7e8f8-8031-463c-8267-255938aad822\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://702c8093144fded24e70730efd8a26f65b34777f7b3d8ab2fe18beb6cc6b6547\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cq7tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2q7bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.369296 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee780ee0-8dc6-4f6f-b47b-d5931c2ba853\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://808a4019218d0e8fe18a085a6027b53da0cccddb5ca08e21452014055b3f4e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34d8b0568d521140f52aeef7d7dbd8adf289151877bf4866df15adcf1a690922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34d8b0568d521140f52aeef7d7dbd8adf289151877bf4866df15adcf1a690922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.380973 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b15ee151e208d1286f536eb3f1b18663e89a934d35772926137fe98f9ff7c508\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.390062 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.390091 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.390100 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.390133 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.390147 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:39Z","lastTransitionTime":"2025-10-04T04:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.399290 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.414698 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"984bf56e-d83a-4206-a887-116d728163d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:28Z\\\",\\\"message\\\":\\\"vices.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"192.168.126.11\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string(nil), Groups:[]string(nil)}}\\\\nI1004 04:35:28.235805 6526 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 04:35:28.235861 6526 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:35:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sfjxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hqg47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.425276 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.434771 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d6616b46a7c52b6cda3bfa8615e97afb08eea31b9f6240ec13925921f24ebc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.444990 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.453717 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df9f335e-3c50-4cbf-ad47-491eb2486de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5df0e293442c2f219c0a3d18f611a0cee4f32e62d25bac48da382f9572c6006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pr2vw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pl4jq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.463990 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sj5hq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31197ccb-a3f9-4130-bca4-ca1c3d3ee065\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:35:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:35:22Z\\\",\\\"message\\\":\\\"2025-10-04T04:34:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a\\\\n2025-10-04T04:34:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_90a33bac-9b3c-4048-a853-d707e3aa840a to /host/opt/cni/bin/\\\\n2025-10-04T04:34:37Z [verbose] multus-daemon started\\\\n2025-10-04T04:34:37Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:35:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w8b6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sj5hq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.477321 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea7b46-6e25-4b35-a516-687dc19bd9ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7515bccd806397b3be4b425fed6b8455f2977d6bf5cb9950e353e8c2153a370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://781b72ceffa508009841f2116377ffd5bca1e9994223ae1e2c0821b298f458af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026695eec5883a503a26e388fef2276c95e5b7a5c261361da52cfe7e22cfed01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90bdcd6150b9321e1c9a8f28d08594c74a0f8aea1d0b66ec8025fb2bd87803b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.496440 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.496478 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.496488 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.496519 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.496529 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:39Z","lastTransitionTime":"2025-10-04T04:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.509115 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e9e342ca5a4ecd744f304e8c2412c154133c40ea7701a185fff3889892dca5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1268ae8db866bf5f047bfdf9016b1a1014eeaf0ca721fb4fc027c2a93a2c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.528437 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2sdsl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1f25b10-0668-41ff-9197-c1fed6995ca6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a84d662c13571547ff0aa68527435a8e857eab2762defd44bba0ad263a501943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjc4f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:29Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2sdsl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.542142 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02c08151-21fd-4ed4-abd7-3ec7e427df35\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba4895d2f323310a99038b7e02a52ff4b77106df50acdf369eb8aaecfd3317c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf0a880f27e6d34fbd878b003b0be3878e9b676eaa358cd99f5e1c2f650dfb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec02e5df9a6061bb3a6648b6886734895f3ff1def41ee4a44aaf6e5a1f1443a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce42c859aae90ef9700cd5057d94707f0d526cf74dbc1af564feab283479746c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7f4002b9e3553298b247a0719a57d3fbc80b99857dc6e33a13280b91a0e1bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffea85f845164938277b5c3a085ef1b584ebc14e3c7a6b1e6ab7f619a7d29a83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9157883f52aad88a6556bc59c7b456be47bf61b26d9994ec56b6c8f3cfd4c88a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpn2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kmr4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.551202 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"331e96fa-b5b5-484d-a2d7-bdbfa6da8ea1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a87c14c6aba75a280bdbd9cc3e05a0ca47cb63ec342a0ebb46b9480dc44d2ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://222f1a424db358bea12d08c165bda9a58174140201c9e79d9cdfae2932a84f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5f6l6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fqfm7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.559890 4575 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9nvms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"026aaa63-845a-45bb-bc23-7b7c3024eead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9ct29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:34:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9nvms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:35:39Z is after 2025-08-24T17:21:41Z" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.598748 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.598787 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.598797 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.598814 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.598825 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:39Z","lastTransitionTime":"2025-10-04T04:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.701225 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.701446 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.701508 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.701706 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.701779 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:39Z","lastTransitionTime":"2025-10-04T04:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.803892 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.803935 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.803943 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.803958 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.803970 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:39Z","lastTransitionTime":"2025-10-04T04:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.906197 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.906229 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.906238 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.906251 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:39 crc kubenswrapper[4575]: I1004 04:35:39.906261 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:39Z","lastTransitionTime":"2025-10-04T04:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.008041 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.008709 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.008734 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.008751 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.008762 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.114250 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.114323 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.114333 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.114348 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.114361 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.217365 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.217412 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.217421 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.217435 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.217444 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.309118 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:40 crc kubenswrapper[4575]: E1004 04:35:40.309437 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.320213 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.320270 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.320283 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.320299 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.320312 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.423214 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.423244 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.423258 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.423285 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.423299 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.525558 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.525604 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.525612 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.525624 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.525633 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.627700 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.627739 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.627750 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.627764 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.627773 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.730787 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.730843 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.730853 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.730870 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.730884 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.833389 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.833426 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.833435 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.833448 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.833456 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.935751 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.935801 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.935811 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.935828 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:40 crc kubenswrapper[4575]: I1004 04:35:40.935839 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:40Z","lastTransitionTime":"2025-10-04T04:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.038125 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.038738 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.038840 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.038943 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.039034 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.141185 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.141217 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.141228 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.141241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.141251 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.243512 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.243582 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.243612 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.243631 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.243641 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.309358 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.309402 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:41 crc kubenswrapper[4575]: E1004 04:35:41.309478 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.309358 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:41 crc kubenswrapper[4575]: E1004 04:35:41.309565 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:41 crc kubenswrapper[4575]: E1004 04:35:41.309661 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.346260 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.346291 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.346299 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.346311 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.346321 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.448263 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.448298 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.448308 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.448323 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.448365 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.550687 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.550723 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.550731 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.550749 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.550762 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.653544 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.653571 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.653582 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.653972 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.653983 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.755375 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.755414 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.755425 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.755440 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.755451 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.857391 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.857423 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.857431 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.857447 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.857458 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.960124 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.960160 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.960171 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.960185 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:41 crc kubenswrapper[4575]: I1004 04:35:41.960196 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:41Z","lastTransitionTime":"2025-10-04T04:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.062971 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.063012 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.063020 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.063033 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.063044 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.166096 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.166132 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.166141 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.166153 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.166163 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.267956 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.268001 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.268009 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.268022 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.268033 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.309480 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:42 crc kubenswrapper[4575]: E1004 04:35:42.309621 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.310175 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:35:42 crc kubenswrapper[4575]: E1004 04:35:42.310328 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.370139 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.370425 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.370566 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.370694 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.370798 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.473652 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.473753 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.473763 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.473794 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.473805 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.575714 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.575772 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.575790 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.575807 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.575816 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.678094 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.678135 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.678144 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.678158 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.678167 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.780660 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.780699 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.780710 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.780725 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.780737 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.883026 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.883066 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.883076 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.883090 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.883100 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.985425 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.985744 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.985823 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.985895 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:42 crc kubenswrapper[4575]: I1004 04:35:42.985973 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:42Z","lastTransitionTime":"2025-10-04T04:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.088430 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.089268 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.089304 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.089319 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.089329 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:43Z","lastTransitionTime":"2025-10-04T04:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.192154 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.192185 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.192192 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.192205 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.192215 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:43Z","lastTransitionTime":"2025-10-04T04:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.294516 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.294544 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.294552 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.294564 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.294573 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:43Z","lastTransitionTime":"2025-10-04T04:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.309840 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.310275 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:43 crc kubenswrapper[4575]: E1004 04:35:43.310433 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.310456 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:43 crc kubenswrapper[4575]: E1004 04:35:43.310578 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:43 crc kubenswrapper[4575]: E1004 04:35:43.311189 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.397867 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.397908 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.397919 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.397935 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.397949 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:43Z","lastTransitionTime":"2025-10-04T04:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.510105 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.510138 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.510147 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.510159 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.510169 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:43Z","lastTransitionTime":"2025-10-04T04:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.612446 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.612494 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.612505 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.612524 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.612539 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:43Z","lastTransitionTime":"2025-10-04T04:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.714739 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.714993 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.715101 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.715216 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.715292 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:43Z","lastTransitionTime":"2025-10-04T04:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.818516 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.818577 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.818613 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.818631 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.818644 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:43Z","lastTransitionTime":"2025-10-04T04:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.922130 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.922227 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.922241 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.922266 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:43 crc kubenswrapper[4575]: I1004 04:35:43.922281 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:43Z","lastTransitionTime":"2025-10-04T04:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.024250 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.024632 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.024645 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.024684 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.024698 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.127497 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.127537 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.127564 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.127604 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.127615 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.229182 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.229215 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.229225 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.229238 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.229247 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.309429 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:44 crc kubenswrapper[4575]: E1004 04:35:44.309979 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.331458 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.331519 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.331530 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.331544 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.331553 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.434259 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.434300 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.434312 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.434328 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.434339 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.536670 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.536736 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.536748 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.536761 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.536771 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.639536 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.639609 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.639622 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.639636 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.639649 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.741414 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.741458 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.741471 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.741486 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.741496 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.843777 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.844128 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.844247 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.844449 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.844654 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.947023 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.947071 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.947080 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.947094 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:44 crc kubenswrapper[4575]: I1004 04:35:44.947104 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:44Z","lastTransitionTime":"2025-10-04T04:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.050178 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.050380 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.050440 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.050524 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.050606 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.152901 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.152937 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.152945 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.152959 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.152969 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.255292 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.255341 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.255355 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.255374 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.255388 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.309369 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.309444 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:45 crc kubenswrapper[4575]: E1004 04:35:45.309555 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.309793 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:45 crc kubenswrapper[4575]: E1004 04:35:45.309845 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:45 crc kubenswrapper[4575]: E1004 04:35:45.309947 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.357619 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.357669 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.357678 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.357690 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.357700 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.460783 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.460831 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.460842 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.460858 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.460868 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.564181 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.564238 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.564250 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.564267 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.564280 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.666460 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.666753 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.666873 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.666983 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.667068 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.769294 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.769338 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.769350 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.769365 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.769375 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.872013 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.872084 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.872106 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.872135 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.872155 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.973874 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.973929 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.973946 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.973971 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:45 crc kubenswrapper[4575]: I1004 04:35:45.973988 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:45Z","lastTransitionTime":"2025-10-04T04:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.075866 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.075901 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.075910 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.075923 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.075932 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:46Z","lastTransitionTime":"2025-10-04T04:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.177927 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.177974 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.177987 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.178004 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.178019 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:46Z","lastTransitionTime":"2025-10-04T04:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.280479 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.280527 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.280539 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.280554 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.280565 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:46Z","lastTransitionTime":"2025-10-04T04:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.309153 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:46 crc kubenswrapper[4575]: E1004 04:35:46.309296 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.383156 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.383185 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.383194 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.383206 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.383215 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:46Z","lastTransitionTime":"2025-10-04T04:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.485607 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.485669 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.485684 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.485726 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.485738 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:46Z","lastTransitionTime":"2025-10-04T04:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.563193 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.563236 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.563248 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.563262 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.563272 4575 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:35:46Z","lastTransitionTime":"2025-10-04T04:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.603185 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw"] Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.603542 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.606124 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.606169 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.606695 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.607068 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.680857 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podStartSLOduration=77.68083949 podStartE2EDuration="1m17.68083949s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.666802522 +0000 UTC m=+97.995361336" watchObservedRunningTime="2025-10-04 04:35:46.68083949 +0000 UTC m=+98.009398294" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.694581 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-sj5hq" podStartSLOduration=77.694565729 podStartE2EDuration="1m17.694565729s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.68119766 +0000 UTC m=+98.009756474" watchObservedRunningTime="2025-10-04 04:35:46.694565729 +0000 UTC m=+98.023124553" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.723980 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=77.723961311 podStartE2EDuration="1m17.723961311s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.709694477 +0000 UTC m=+98.038253301" watchObservedRunningTime="2025-10-04 04:35:46.723961311 +0000 UTC m=+98.052520135" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.753470 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-kmr4x" podStartSLOduration=77.753453057 podStartE2EDuration="1m17.753453057s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.753064286 +0000 UTC m=+98.081623110" watchObservedRunningTime="2025-10-04 04:35:46.753453057 +0000 UTC m=+98.082011871" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.753684 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-2sdsl" podStartSLOduration=77.753680743 podStartE2EDuration="1m17.753680743s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.735541609 +0000 UTC m=+98.064100423" watchObservedRunningTime="2025-10-04 04:35:46.753680743 +0000 UTC m=+98.082239557" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.767188 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fqfm7" podStartSLOduration=76.767173225 podStartE2EDuration="1m16.767173225s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.766572128 +0000 UTC m=+98.095130952" watchObservedRunningTime="2025-10-04 04:35:46.767173225 +0000 UTC m=+98.095732039" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.793031 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=12.793014367 podStartE2EDuration="12.793014367s" podCreationTimestamp="2025-10-04 04:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.790775664 +0000 UTC m=+98.119334488" watchObservedRunningTime="2025-10-04 04:35:46.793014367 +0000 UTC m=+98.121573181" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.802752 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccc813ac-9412-4ea7-9e48-ae7083d5038c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.802848 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ccc813ac-9412-4ea7-9e48-ae7083d5038c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.802893 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccc813ac-9412-4ea7-9e48-ae7083d5038c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.803010 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ccc813ac-9412-4ea7-9e48-ae7083d5038c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.803068 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ccc813ac-9412-4ea7-9e48-ae7083d5038c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.807848 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.807830557 podStartE2EDuration="1m17.807830557s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.807453766 +0000 UTC m=+98.136012600" watchObservedRunningTime="2025-10-04 04:35:46.807830557 +0000 UTC m=+98.136389371" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.819091 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.819060665 podStartE2EDuration="48.819060665s" podCreationTimestamp="2025-10-04 04:34:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.818232502 +0000 UTC m=+98.146791316" watchObservedRunningTime="2025-10-04 04:35:46.819060665 +0000 UTC m=+98.147619479" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.828654 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2q7bk" podStartSLOduration=77.828639096 podStartE2EDuration="1m17.828639096s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.82842774 +0000 UTC m=+98.156986554" watchObservedRunningTime="2025-10-04 04:35:46.828639096 +0000 UTC m=+98.157197910" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.851436 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=15.851400511 podStartE2EDuration="15.851400511s" podCreationTimestamp="2025-10-04 04:35:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:46.839540655 +0000 UTC m=+98.168099459" watchObservedRunningTime="2025-10-04 04:35:46.851400511 +0000 UTC m=+98.179959325" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.878950 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.879804 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:35:46 crc kubenswrapper[4575]: E1004 04:35:46.879982 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.904489 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ccc813ac-9412-4ea7-9e48-ae7083d5038c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.904539 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ccc813ac-9412-4ea7-9e48-ae7083d5038c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.904574 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccc813ac-9412-4ea7-9e48-ae7083d5038c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.904616 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ccc813ac-9412-4ea7-9e48-ae7083d5038c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.904629 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ccc813ac-9412-4ea7-9e48-ae7083d5038c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.904650 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccc813ac-9412-4ea7-9e48-ae7083d5038c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.904997 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ccc813ac-9412-4ea7-9e48-ae7083d5038c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.905657 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ccc813ac-9412-4ea7-9e48-ae7083d5038c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.913359 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccc813ac-9412-4ea7-9e48-ae7083d5038c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:46 crc kubenswrapper[4575]: I1004 04:35:46.921208 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccc813ac-9412-4ea7-9e48-ae7083d5038c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-94vsw\" (UID: \"ccc813ac-9412-4ea7-9e48-ae7083d5038c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:47 crc kubenswrapper[4575]: I1004 04:35:47.216561 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" Oct 04 04:35:47 crc kubenswrapper[4575]: I1004 04:35:47.309515 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:47 crc kubenswrapper[4575]: I1004 04:35:47.309525 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:47 crc kubenswrapper[4575]: E1004 04:35:47.309627 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:47 crc kubenswrapper[4575]: I1004 04:35:47.309675 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:47 crc kubenswrapper[4575]: E1004 04:35:47.309772 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:47 crc kubenswrapper[4575]: E1004 04:35:47.309867 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:47 crc kubenswrapper[4575]: I1004 04:35:47.768981 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" event={"ID":"ccc813ac-9412-4ea7-9e48-ae7083d5038c","Type":"ContainerStarted","Data":"695508184c4b6179f5f1b1611f6e18cb607aea30237770cf11bb375dc5f10fd8"} Oct 04 04:35:47 crc kubenswrapper[4575]: I1004 04:35:47.769029 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" event={"ID":"ccc813ac-9412-4ea7-9e48-ae7083d5038c","Type":"ContainerStarted","Data":"61bf5d6fbf42b2e079574af460637952a525c8f3ee96088eddb7e24b27fe56ca"} Oct 04 04:35:47 crc kubenswrapper[4575]: I1004 04:35:47.783419 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-94vsw" podStartSLOduration=78.78339766 podStartE2EDuration="1m18.78339766s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:47.781120685 +0000 UTC m=+99.109679519" watchObservedRunningTime="2025-10-04 04:35:47.78339766 +0000 UTC m=+99.111956474" Oct 04 04:35:47 crc kubenswrapper[4575]: I1004 04:35:47.813338 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:47 crc kubenswrapper[4575]: E1004 04:35:47.813522 4575 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:35:47 crc kubenswrapper[4575]: E1004 04:35:47.813628 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs podName:026aaa63-845a-45bb-bc23-7b7c3024eead nodeName:}" failed. No retries permitted until 2025-10-04 04:36:51.813576445 +0000 UTC m=+163.142135269 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs") pod "network-metrics-daemon-9nvms" (UID: "026aaa63-845a-45bb-bc23-7b7c3024eead") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:35:48 crc kubenswrapper[4575]: I1004 04:35:48.309856 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:48 crc kubenswrapper[4575]: E1004 04:35:48.310188 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:49 crc kubenswrapper[4575]: I1004 04:35:49.309261 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:49 crc kubenswrapper[4575]: I1004 04:35:49.309289 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:49 crc kubenswrapper[4575]: E1004 04:35:49.310247 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:49 crc kubenswrapper[4575]: E1004 04:35:49.310536 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:49 crc kubenswrapper[4575]: I1004 04:35:49.310882 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:49 crc kubenswrapper[4575]: E1004 04:35:49.311055 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:50 crc kubenswrapper[4575]: I1004 04:35:50.309751 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:50 crc kubenswrapper[4575]: E1004 04:35:50.309881 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:51 crc kubenswrapper[4575]: I1004 04:35:51.309396 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:51 crc kubenswrapper[4575]: I1004 04:35:51.309396 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:51 crc kubenswrapper[4575]: I1004 04:35:51.309490 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:51 crc kubenswrapper[4575]: E1004 04:35:51.309674 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:51 crc kubenswrapper[4575]: E1004 04:35:51.309782 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:51 crc kubenswrapper[4575]: E1004 04:35:51.310219 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:52 crc kubenswrapper[4575]: I1004 04:35:52.309020 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:52 crc kubenswrapper[4575]: E1004 04:35:52.309383 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:53 crc kubenswrapper[4575]: I1004 04:35:53.309627 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:53 crc kubenswrapper[4575]: E1004 04:35:53.309769 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:53 crc kubenswrapper[4575]: I1004 04:35:53.309822 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:53 crc kubenswrapper[4575]: E1004 04:35:53.309884 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:53 crc kubenswrapper[4575]: I1004 04:35:53.310025 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:53 crc kubenswrapper[4575]: E1004 04:35:53.310081 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:54 crc kubenswrapper[4575]: I1004 04:35:54.309632 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:54 crc kubenswrapper[4575]: E1004 04:35:54.309890 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:55 crc kubenswrapper[4575]: I1004 04:35:55.309504 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:55 crc kubenswrapper[4575]: E1004 04:35:55.309663 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:55 crc kubenswrapper[4575]: I1004 04:35:55.309695 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:55 crc kubenswrapper[4575]: E1004 04:35:55.309817 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:55 crc kubenswrapper[4575]: I1004 04:35:55.309522 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:55 crc kubenswrapper[4575]: E1004 04:35:55.309900 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:56 crc kubenswrapper[4575]: I1004 04:35:56.309252 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:56 crc kubenswrapper[4575]: E1004 04:35:56.309390 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:57 crc kubenswrapper[4575]: I1004 04:35:57.309111 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:57 crc kubenswrapper[4575]: I1004 04:35:57.309227 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:57 crc kubenswrapper[4575]: E1004 04:35:57.309424 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:57 crc kubenswrapper[4575]: I1004 04:35:57.309497 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:57 crc kubenswrapper[4575]: E1004 04:35:57.309549 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:57 crc kubenswrapper[4575]: E1004 04:35:57.309245 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:58 crc kubenswrapper[4575]: I1004 04:35:58.309070 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:58 crc kubenswrapper[4575]: E1004 04:35:58.309260 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:59 crc kubenswrapper[4575]: I1004 04:35:59.309144 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:35:59 crc kubenswrapper[4575]: I1004 04:35:59.309198 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:59 crc kubenswrapper[4575]: I1004 04:35:59.310756 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:59 crc kubenswrapper[4575]: E1004 04:35:59.310821 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:59 crc kubenswrapper[4575]: E1004 04:35:59.310882 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:59 crc kubenswrapper[4575]: E1004 04:35:59.311329 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:35:59 crc kubenswrapper[4575]: I1004 04:35:59.311643 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:35:59 crc kubenswrapper[4575]: E1004 04:35:59.311851 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hqg47_openshift-ovn-kubernetes(984bf56e-d83a-4206-a887-116d728163d1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" Oct 04 04:36:00 crc kubenswrapper[4575]: I1004 04:36:00.309503 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:00 crc kubenswrapper[4575]: E1004 04:36:00.310200 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:01 crc kubenswrapper[4575]: I1004 04:36:01.309646 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:01 crc kubenswrapper[4575]: E1004 04:36:01.309783 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:01 crc kubenswrapper[4575]: I1004 04:36:01.309818 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:01 crc kubenswrapper[4575]: I1004 04:36:01.309884 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:01 crc kubenswrapper[4575]: E1004 04:36:01.309960 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:01 crc kubenswrapper[4575]: E1004 04:36:01.310021 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:02 crc kubenswrapper[4575]: I1004 04:36:02.309580 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:02 crc kubenswrapper[4575]: E1004 04:36:02.309766 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:03 crc kubenswrapper[4575]: I1004 04:36:03.309267 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:03 crc kubenswrapper[4575]: I1004 04:36:03.309267 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:03 crc kubenswrapper[4575]: E1004 04:36:03.309412 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:03 crc kubenswrapper[4575]: E1004 04:36:03.309474 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:03 crc kubenswrapper[4575]: I1004 04:36:03.309288 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:03 crc kubenswrapper[4575]: E1004 04:36:03.309552 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:04 crc kubenswrapper[4575]: I1004 04:36:04.309159 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:04 crc kubenswrapper[4575]: E1004 04:36:04.309263 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:05 crc kubenswrapper[4575]: I1004 04:36:05.308915 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:05 crc kubenswrapper[4575]: I1004 04:36:05.309004 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:05 crc kubenswrapper[4575]: E1004 04:36:05.309060 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:05 crc kubenswrapper[4575]: E1004 04:36:05.309124 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:05 crc kubenswrapper[4575]: I1004 04:36:05.309212 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:05 crc kubenswrapper[4575]: E1004 04:36:05.309259 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:06 crc kubenswrapper[4575]: I1004 04:36:06.309804 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:06 crc kubenswrapper[4575]: E1004 04:36:06.310009 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:07 crc kubenswrapper[4575]: I1004 04:36:07.309642 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:07 crc kubenswrapper[4575]: I1004 04:36:07.309669 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:07 crc kubenswrapper[4575]: I1004 04:36:07.309688 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:07 crc kubenswrapper[4575]: E1004 04:36:07.309811 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:07 crc kubenswrapper[4575]: E1004 04:36:07.309947 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:07 crc kubenswrapper[4575]: E1004 04:36:07.310022 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:08 crc kubenswrapper[4575]: I1004 04:36:08.308862 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:08 crc kubenswrapper[4575]: E1004 04:36:08.308990 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:09 crc kubenswrapper[4575]: I1004 04:36:09.309410 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:09 crc kubenswrapper[4575]: I1004 04:36:09.310702 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:09 crc kubenswrapper[4575]: E1004 04:36:09.310720 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:09 crc kubenswrapper[4575]: I1004 04:36:09.310763 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:09 crc kubenswrapper[4575]: E1004 04:36:09.310871 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:09 crc kubenswrapper[4575]: E1004 04:36:09.311196 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:09 crc kubenswrapper[4575]: E1004 04:36:09.346570 4575 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 04 04:36:09 crc kubenswrapper[4575]: E1004 04:36:09.388848 4575 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 04:36:09 crc kubenswrapper[4575]: I1004 04:36:09.835736 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sj5hq_31197ccb-a3f9-4130-bca4-ca1c3d3ee065/kube-multus/1.log" Oct 04 04:36:09 crc kubenswrapper[4575]: I1004 04:36:09.836238 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sj5hq_31197ccb-a3f9-4130-bca4-ca1c3d3ee065/kube-multus/0.log" Oct 04 04:36:09 crc kubenswrapper[4575]: I1004 04:36:09.836290 4575 generic.go:334] "Generic (PLEG): container finished" podID="31197ccb-a3f9-4130-bca4-ca1c3d3ee065" containerID="ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f" exitCode=1 Oct 04 04:36:09 crc kubenswrapper[4575]: I1004 04:36:09.836322 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sj5hq" event={"ID":"31197ccb-a3f9-4130-bca4-ca1c3d3ee065","Type":"ContainerDied","Data":"ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f"} Oct 04 04:36:09 crc kubenswrapper[4575]: I1004 04:36:09.836356 4575 scope.go:117] "RemoveContainer" containerID="adc544196fcf4b446334396319aba8675ad46e882e1a8ef57f0e9bec94526649" Oct 04 04:36:09 crc kubenswrapper[4575]: I1004 04:36:09.837183 4575 scope.go:117] "RemoveContainer" containerID="ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f" Oct 04 04:36:09 crc kubenswrapper[4575]: E1004 04:36:09.837551 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-sj5hq_openshift-multus(31197ccb-a3f9-4130-bca4-ca1c3d3ee065)\"" pod="openshift-multus/multus-sj5hq" podUID="31197ccb-a3f9-4130-bca4-ca1c3d3ee065" Oct 04 04:36:10 crc kubenswrapper[4575]: I1004 04:36:10.309905 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:10 crc kubenswrapper[4575]: E1004 04:36:10.310067 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:10 crc kubenswrapper[4575]: I1004 04:36:10.840904 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sj5hq_31197ccb-a3f9-4130-bca4-ca1c3d3ee065/kube-multus/1.log" Oct 04 04:36:11 crc kubenswrapper[4575]: I1004 04:36:11.309562 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:11 crc kubenswrapper[4575]: I1004 04:36:11.309569 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:11 crc kubenswrapper[4575]: E1004 04:36:11.309702 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:11 crc kubenswrapper[4575]: E1004 04:36:11.309781 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:11 crc kubenswrapper[4575]: I1004 04:36:11.309606 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:11 crc kubenswrapper[4575]: E1004 04:36:11.309856 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:12 crc kubenswrapper[4575]: I1004 04:36:12.309073 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:12 crc kubenswrapper[4575]: E1004 04:36:12.309198 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:13 crc kubenswrapper[4575]: I1004 04:36:13.309682 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:13 crc kubenswrapper[4575]: I1004 04:36:13.309721 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:13 crc kubenswrapper[4575]: I1004 04:36:13.309700 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:13 crc kubenswrapper[4575]: E1004 04:36:13.309856 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:13 crc kubenswrapper[4575]: E1004 04:36:13.309932 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:13 crc kubenswrapper[4575]: E1004 04:36:13.310075 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:14 crc kubenswrapper[4575]: I1004 04:36:14.308849 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:14 crc kubenswrapper[4575]: E1004 04:36:14.308961 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:14 crc kubenswrapper[4575]: I1004 04:36:14.310028 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:36:14 crc kubenswrapper[4575]: E1004 04:36:14.391142 4575 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 04:36:14 crc kubenswrapper[4575]: I1004 04:36:14.855126 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/3.log" Oct 04 04:36:14 crc kubenswrapper[4575]: I1004 04:36:14.856911 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerStarted","Data":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} Oct 04 04:36:15 crc kubenswrapper[4575]: I1004 04:36:15.187989 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9nvms"] Oct 04 04:36:15 crc kubenswrapper[4575]: I1004 04:36:15.188110 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:15 crc kubenswrapper[4575]: E1004 04:36:15.188210 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:15 crc kubenswrapper[4575]: I1004 04:36:15.309842 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:15 crc kubenswrapper[4575]: I1004 04:36:15.309931 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:15 crc kubenswrapper[4575]: E1004 04:36:15.310028 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:15 crc kubenswrapper[4575]: E1004 04:36:15.310205 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:15 crc kubenswrapper[4575]: I1004 04:36:15.859917 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:36:16 crc kubenswrapper[4575]: I1004 04:36:16.309870 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:16 crc kubenswrapper[4575]: E1004 04:36:16.310048 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:17 crc kubenswrapper[4575]: I1004 04:36:17.309226 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:17 crc kubenswrapper[4575]: E1004 04:36:17.309344 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:17 crc kubenswrapper[4575]: I1004 04:36:17.309225 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:17 crc kubenswrapper[4575]: I1004 04:36:17.309227 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:17 crc kubenswrapper[4575]: E1004 04:36:17.309515 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:17 crc kubenswrapper[4575]: E1004 04:36:17.309537 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:18 crc kubenswrapper[4575]: I1004 04:36:18.309071 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:18 crc kubenswrapper[4575]: E1004 04:36:18.309199 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:19 crc kubenswrapper[4575]: I1004 04:36:19.309190 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:19 crc kubenswrapper[4575]: I1004 04:36:19.309209 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:19 crc kubenswrapper[4575]: I1004 04:36:19.310150 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:19 crc kubenswrapper[4575]: E1004 04:36:19.310149 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:19 crc kubenswrapper[4575]: E1004 04:36:19.310304 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:19 crc kubenswrapper[4575]: E1004 04:36:19.310666 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:19 crc kubenswrapper[4575]: E1004 04:36:19.391962 4575 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 04:36:20 crc kubenswrapper[4575]: I1004 04:36:20.309011 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:20 crc kubenswrapper[4575]: E1004 04:36:20.309132 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:21 crc kubenswrapper[4575]: I1004 04:36:21.309376 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:21 crc kubenswrapper[4575]: E1004 04:36:21.309886 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:21 crc kubenswrapper[4575]: I1004 04:36:21.309481 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:21 crc kubenswrapper[4575]: E1004 04:36:21.309976 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:21 crc kubenswrapper[4575]: I1004 04:36:21.309413 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:21 crc kubenswrapper[4575]: E1004 04:36:21.310035 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:22 crc kubenswrapper[4575]: I1004 04:36:22.309714 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:22 crc kubenswrapper[4575]: E1004 04:36:22.309802 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:22 crc kubenswrapper[4575]: I1004 04:36:22.309974 4575 scope.go:117] "RemoveContainer" containerID="ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f" Oct 04 04:36:22 crc kubenswrapper[4575]: I1004 04:36:22.330696 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podStartSLOduration=113.330676374 podStartE2EDuration="1m53.330676374s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:15.885849719 +0000 UTC m=+127.214408543" watchObservedRunningTime="2025-10-04 04:36:22.330676374 +0000 UTC m=+133.659235198" Oct 04 04:36:22 crc kubenswrapper[4575]: I1004 04:36:22.880678 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sj5hq_31197ccb-a3f9-4130-bca4-ca1c3d3ee065/kube-multus/1.log" Oct 04 04:36:22 crc kubenswrapper[4575]: I1004 04:36:22.880722 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sj5hq" event={"ID":"31197ccb-a3f9-4130-bca4-ca1c3d3ee065","Type":"ContainerStarted","Data":"c3913b32c9fe101cb5dd4815993d5e8747107d47f003f32bdcade9542f016a66"} Oct 04 04:36:23 crc kubenswrapper[4575]: I1004 04:36:23.309396 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:23 crc kubenswrapper[4575]: E1004 04:36:23.309827 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9nvms" podUID="026aaa63-845a-45bb-bc23-7b7c3024eead" Oct 04 04:36:23 crc kubenswrapper[4575]: I1004 04:36:23.309566 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:23 crc kubenswrapper[4575]: E1004 04:36:23.310635 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:36:23 crc kubenswrapper[4575]: I1004 04:36:23.309516 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:23 crc kubenswrapper[4575]: E1004 04:36:23.310886 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:36:24 crc kubenswrapper[4575]: I1004 04:36:24.309178 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:24 crc kubenswrapper[4575]: E1004 04:36:24.309287 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:36:25 crc kubenswrapper[4575]: I1004 04:36:25.309774 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:25 crc kubenswrapper[4575]: I1004 04:36:25.309909 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:25 crc kubenswrapper[4575]: I1004 04:36:25.309782 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:25 crc kubenswrapper[4575]: I1004 04:36:25.314652 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 04 04:36:25 crc kubenswrapper[4575]: I1004 04:36:25.315040 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 04 04:36:25 crc kubenswrapper[4575]: I1004 04:36:25.315053 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 04 04:36:25 crc kubenswrapper[4575]: I1004 04:36:25.315054 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 04 04:36:25 crc kubenswrapper[4575]: I1004 04:36:25.315086 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 04 04:36:25 crc kubenswrapper[4575]: I1004 04:36:25.315150 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 04 04:36:26 crc kubenswrapper[4575]: I1004 04:36:26.309246 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:26 crc kubenswrapper[4575]: I1004 04:36:26.992520 4575 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.023246 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.023784 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.024348 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q9ssb"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.024884 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.025531 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flrql"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.026024 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.026428 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rdr7j"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.026832 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.032377 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mmjn9"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.032745 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.033866 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.034367 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.034397 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.037168 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: W1004 04:36:27.037715 4575 reflector.go:561] object-"openshift-authentication-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Oct 04 04:36:27 crc kubenswrapper[4575]: E1004 04:36:27.037782 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:36:27 crc kubenswrapper[4575]: W1004 04:36:27.039044 4575 reflector.go:561] object-"openshift-authentication-operator"/"service-ca-bundle": failed to list *v1.ConfigMap: configmaps "service-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Oct 04 04:36:27 crc kubenswrapper[4575]: E1004 04:36:27.039082 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"service-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"service-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:36:27 crc kubenswrapper[4575]: W1004 04:36:27.039168 4575 reflector.go:561] object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj": failed to list *v1.Secret: secrets "authentication-operator-dockercfg-mz9bj" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Oct 04 04:36:27 crc kubenswrapper[4575]: E1004 04:36:27.039212 4575 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-mz9bj\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"authentication-operator-dockercfg-mz9bj\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.041078 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.045166 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.060667 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.061313 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.061501 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.062124 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.062153 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.061502 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.062316 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.062437 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.062578 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.062803 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.062815 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.066001 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.067252 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.067741 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.068226 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.068387 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.068535 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.068699 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.068764 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.068881 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.068935 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069094 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069134 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069175 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069245 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.068703 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069439 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069610 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069705 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069850 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069918 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069935 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.069984 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.070028 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.070112 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.071406 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.071424 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wk2s"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.071917 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.072781 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.074017 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.075894 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.076017 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-vgclk"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.078281 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.078632 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.082686 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.082700 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.083064 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.083111 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.083277 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.083455 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.084972 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2q455"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.085692 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.085871 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.086827 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.086968 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.087019 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.087238 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.087242 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.090806 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-9r5mx"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.091298 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f7lqb"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.091629 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.091916 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9r5mx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.091997 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.092078 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.095683 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.096210 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q9ssb"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.096303 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.099318 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.101025 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.101450 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.101790 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.102063 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.102224 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131196 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cdkp\" (UniqueName: \"kubernetes.io/projected/c72540e8-b2dd-4377-8a17-717fcb2a05a2-kube-api-access-8cdkp\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131232 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-config\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131260 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hgrg\" (UniqueName: \"kubernetes.io/projected/14d2bb1f-8844-48cd-ba55-0782b444064e-kube-api-access-6hgrg\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131281 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-client-ca\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131301 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14d2bb1f-8844-48cd-ba55-0782b444064e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131322 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/021e5871-36e8-48d3-92de-e504a9429d9e-serving-cert\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131355 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a04cd83-dbd7-4076-858a-ff36748ed2d1-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131380 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6d607e6e-db54-4e24-aafa-8fa7c16c949b-images\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131400 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14d2bb1f-8844-48cd-ba55-0782b444064e-audit-policies\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131431 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhtvr\" (UniqueName: \"kubernetes.io/projected/6d607e6e-db54-4e24-aafa-8fa7c16c949b-kube-api-access-qhtvr\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131454 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a04cd83-dbd7-4076-858a-ff36748ed2d1-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131474 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sswjt\" (UniqueName: \"kubernetes.io/projected/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-kube-api-access-sswjt\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131496 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-etcd-serving-ca\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131517 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/14d2bb1f-8844-48cd-ba55-0782b444064e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131535 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c72540e8-b2dd-4377-8a17-717fcb2a05a2-serving-cert\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131551 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c59hf\" (UniqueName: \"kubernetes.io/projected/88d50f89-59b7-4772-8f49-7f7bfcc787ff-kube-api-access-c59hf\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131569 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131605 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d607e6e-db54-4e24-aafa-8fa7c16c949b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131624 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14d2bb1f-8844-48cd-ba55-0782b444064e-audit-dir\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131639 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88d50f89-59b7-4772-8f49-7f7bfcc787ff-audit-dir\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131660 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/88d50f89-59b7-4772-8f49-7f7bfcc787ff-node-pullsecrets\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131679 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-audit\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131704 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/14d2bb1f-8844-48cd-ba55-0782b444064e-encryption-config\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131736 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/88d50f89-59b7-4772-8f49-7f7bfcc787ff-etcd-client\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131757 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-service-ca-bundle\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131784 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hknmw\" (UniqueName: \"kubernetes.io/projected/4a04cd83-dbd7-4076-858a-ff36748ed2d1-kube-api-access-hknmw\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131816 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d607e6e-db54-4e24-aafa-8fa7c16c949b-config\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131832 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131852 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-config\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131872 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-client-ca\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131890 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14d2bb1f-8844-48cd-ba55-0782b444064e-serving-cert\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131905 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88d50f89-59b7-4772-8f49-7f7bfcc787ff-serving-cert\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131922 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-config\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131941 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-config\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131959 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/88d50f89-59b7-4772-8f49-7f7bfcc787ff-encryption-config\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131977 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-serving-cert\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.131991 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-image-import-ca\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.132013 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a04cd83-dbd7-4076-858a-ff36748ed2d1-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.132031 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/14d2bb1f-8844-48cd-ba55-0782b444064e-etcd-client\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.132047 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.132063 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls796\" (UniqueName: \"kubernetes.io/projected/021e5871-36e8-48d3-92de-e504a9429d9e-kube-api-access-ls796\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.133859 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.143618 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.144432 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.144604 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.144447 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.145113 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.145165 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.145442 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.145692 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.145788 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pn9hw"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.145998 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.146543 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.146630 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.146953 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.147147 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.147269 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.147614 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.147977 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.147983 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.148355 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.148544 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.153492 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.153907 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.153957 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154132 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154277 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154303 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154390 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154433 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154567 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154737 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154819 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154881 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.155053 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.155107 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.155143 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.155375 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.154830 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.155761 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.155805 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.155841 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.155912 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.158500 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.158625 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.165806 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.166099 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.166527 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.167854 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.168198 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.168503 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.168693 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.168840 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.168974 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.169115 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.169742 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.169870 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.170378 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.170433 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-bd5sj"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.170761 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.170832 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.171266 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.171358 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.171516 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-prnp7"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.172888 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h5dwl"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.173346 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.173348 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.177025 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.177728 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.178241 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.178513 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.178687 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.178847 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.178903 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.200041 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.200852 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.206198 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.209693 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-n9zm5"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.209967 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.210101 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.210628 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.210964 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.213295 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.214092 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.223031 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.226673 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c54b8"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.227364 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fn6w5"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.227939 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.227966 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.228549 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.228774 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.228785 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.231401 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-b9fb4"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.231981 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.232481 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/88d50f89-59b7-4772-8f49-7f7bfcc787ff-node-pullsecrets\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.232516 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88d50f89-59b7-4772-8f49-7f7bfcc787ff-audit-dir\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.232546 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cdddbf25-7363-4abe-8363-90ed28ff079e-images\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.232569 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.233074 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88d50f89-59b7-4772-8f49-7f7bfcc787ff-audit-dir\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237716 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-audit\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237769 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/14d2bb1f-8844-48cd-ba55-0782b444064e-encryption-config\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237800 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-config\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237821 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237840 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/88d50f89-59b7-4772-8f49-7f7bfcc787ff-etcd-client\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237878 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cdddbf25-7363-4abe-8363-90ed28ff079e-proxy-tls\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237901 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-machine-approver-tls\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237942 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-service-ca-bundle\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237962 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d607e6e-db54-4e24-aafa-8fa7c16c949b-config\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237977 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hknmw\" (UniqueName: \"kubernetes.io/projected/4a04cd83-dbd7-4076-858a-ff36748ed2d1-kube-api-access-hknmw\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.237996 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30854660-94ff-476c-a865-d326c3874db9-config\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.238011 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rdpj\" (UniqueName: \"kubernetes.io/projected/30854660-94ff-476c-a865-d326c3874db9-kube-api-access-8rdpj\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.238050 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.238099 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/88d50f89-59b7-4772-8f49-7f7bfcc787ff-node-pullsecrets\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.238151 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.238172 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-oauth-serving-cert\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.238196 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-config\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.238222 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qrhm\" (UniqueName: \"kubernetes.io/projected/289056ee-b4e0-4eda-b826-5112ba8e54c0-kube-api-access-4qrhm\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtmvv\" (UID: \"289056ee-b4e0-4eda-b826-5112ba8e54c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.238244 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.238716 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-audit\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.239237 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.239484 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-client-ca\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.239528 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-audit-policies\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.239714 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-config\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.240390 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d607e6e-db54-4e24-aafa-8fa7c16c949b-config\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.240936 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14d2bb1f-8844-48cd-ba55-0782b444064e-serving-cert\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.240981 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88d50f89-59b7-4772-8f49-7f7bfcc787ff-serving-cert\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.240998 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-config\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241034 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cdddbf25-7363-4abe-8363-90ed28ff079e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241053 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85lrl\" (UniqueName: \"kubernetes.io/projected/dbe8f63c-7cac-47fb-b961-58c717e7a951-kube-api-access-85lrl\") pod \"dns-operator-744455d44c-pn9hw\" (UID: \"dbe8f63c-7cac-47fb-b961-58c717e7a951\") " pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241075 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241094 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-config\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241111 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mrcs\" (UniqueName: \"kubernetes.io/projected/3322e445-5258-46de-8fda-0c567244c1fa-kube-api-access-7mrcs\") pod \"openshift-apiserver-operator-796bbdcf4f-5v5vj\" (UID: \"3322e445-5258-46de-8fda-0c567244c1fa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241127 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/88d50f89-59b7-4772-8f49-7f7bfcc787ff-encryption-config\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241142 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-serving-cert\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241196 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/205c1018-8396-442c-b265-42ee18193eba-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cdq5t\" (UID: \"205c1018-8396-442c-b265-42ee18193eba\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241214 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241234 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-image-import-ca\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241250 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-trusted-ca-bundle\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241288 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a04cd83-dbd7-4076-858a-ff36748ed2d1-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241305 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3322e445-5258-46de-8fda-0c567244c1fa-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5v5vj\" (UID: \"3322e445-5258-46de-8fda-0c567244c1fa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241321 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-config\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241337 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1510c298-3548-442a-af69-72a237934ef4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2q455\" (UID: \"1510c298-3548-442a-af69-72a237934ef4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241355 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/14d2bb1f-8844-48cd-ba55-0782b444064e-etcd-client\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241398 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241413 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls796\" (UniqueName: \"kubernetes.io/projected/021e5871-36e8-48d3-92de-e504a9429d9e-kube-api-access-ls796\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241431 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t5sd\" (UniqueName: \"kubernetes.io/projected/82bbeaf3-64ae-4676-a365-d75fdf225448-kube-api-access-6t5sd\") pod \"downloads-7954f5f757-9r5mx\" (UID: \"82bbeaf3-64ae-4676-a365-d75fdf225448\") " pod="openshift-console/downloads-7954f5f757-9r5mx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241446 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cdkp\" (UniqueName: \"kubernetes.io/projected/c72540e8-b2dd-4377-8a17-717fcb2a05a2-kube-api-access-8cdkp\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241463 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-config\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241480 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hgrg\" (UniqueName: \"kubernetes.io/projected/14d2bb1f-8844-48cd-ba55-0782b444064e-kube-api-access-6hgrg\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241501 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-client-ca\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241523 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/289056ee-b4e0-4eda-b826-5112ba8e54c0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtmvv\" (UID: \"289056ee-b4e0-4eda-b826-5112ba8e54c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241545 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-serving-cert\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241567 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14d2bb1f-8844-48cd-ba55-0782b444064e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241610 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/289056ee-b4e0-4eda-b826-5112ba8e54c0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtmvv\" (UID: \"289056ee-b4e0-4eda-b826-5112ba8e54c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.241633 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-drt6k\" (UID: \"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.243083 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-config\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.243522 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-config\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.243864 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.244833 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-config\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.245427 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-client-ca\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.246252 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.246279 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.247051 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-image-import-ca\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.247572 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14d2bb1f-8844-48cd-ba55-0782b444064e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.247657 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1-config\") pod \"kube-controller-manager-operator-78b949d7b-drt6k\" (UID: \"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.247693 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.247714 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g45h5\" (UniqueName: \"kubernetes.io/projected/952902f2-2371-4a82-84e7-365978f206f6-kube-api-access-g45h5\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.247730 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jjm6\" (UniqueName: \"kubernetes.io/projected/2fe0947a-c14b-498c-ba57-c5f8733ae76f-kube-api-access-8jjm6\") pod \"control-plane-machine-set-operator-78cbb6b69f-5js5f\" (UID: \"2fe0947a-c14b-498c-ba57-c5f8733ae76f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.247777 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a04cd83-dbd7-4076-858a-ff36748ed2d1-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.247805 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/021e5871-36e8-48d3-92de-e504a9429d9e-serving-cert\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.247821 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.248141 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rdr7j"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.248255 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6d607e6e-db54-4e24-aafa-8fa7c16c949b-images\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.248348 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7lsm\" (UniqueName: \"kubernetes.io/projected/cdddbf25-7363-4abe-8363-90ed28ff079e-kube-api-access-r7lsm\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.248377 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-drt6k\" (UID: \"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.248400 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxkht\" (UniqueName: \"kubernetes.io/projected/1510c298-3548-442a-af69-72a237934ef4-kube-api-access-mxkht\") pod \"openshift-config-operator-7777fb866f-2q455\" (UID: \"1510c298-3548-442a-af69-72a237934ef4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.248651 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flrql"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249034 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3322e445-5258-46de-8fda-0c567244c1fa-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5v5vj\" (UID: \"3322e445-5258-46de-8fda-0c567244c1fa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249067 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54qwr\" (UniqueName: \"kubernetes.io/projected/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-kube-api-access-54qwr\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249322 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-client-ca\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249392 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14d2bb1f-8844-48cd-ba55-0782b444064e-audit-policies\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249430 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/952902f2-2371-4a82-84e7-365978f206f6-audit-dir\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249449 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2fe0947a-c14b-498c-ba57-c5f8733ae76f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5js5f\" (UID: \"2fe0947a-c14b-498c-ba57-c5f8733ae76f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249476 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a04cd83-dbd7-4076-858a-ff36748ed2d1-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249494 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30854660-94ff-476c-a865-d326c3874db9-serving-cert\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249510 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx8vl\" (UniqueName: \"kubernetes.io/projected/8bf9078b-7040-4b98-86d5-1199c124ccb1-kube-api-access-xx8vl\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249530 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249552 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhtvr\" (UniqueName: \"kubernetes.io/projected/6d607e6e-db54-4e24-aafa-8fa7c16c949b-kube-api-access-qhtvr\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249621 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1510c298-3548-442a-af69-72a237934ef4-serving-cert\") pod \"openshift-config-operator-7777fb866f-2q455\" (UID: \"1510c298-3548-442a-af69-72a237934ef4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249640 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-oauth-config\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249656 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249676 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sswjt\" (UniqueName: \"kubernetes.io/projected/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-kube-api-access-sswjt\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249694 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggbsv\" (UniqueName: \"kubernetes.io/projected/205c1018-8396-442c-b265-42ee18193eba-kube-api-access-ggbsv\") pod \"cluster-samples-operator-665b6dd947-cdq5t\" (UID: \"205c1018-8396-442c-b265-42ee18193eba\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249711 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-auth-proxy-config\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249727 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249784 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-etcd-serving-ca\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249805 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c72540e8-b2dd-4377-8a17-717fcb2a05a2-serving-cert\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249821 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30854660-94ff-476c-a865-d326c3874db9-trusted-ca\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249837 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dbe8f63c-7cac-47fb-b961-58c717e7a951-metrics-tls\") pod \"dns-operator-744455d44c-pn9hw\" (UID: \"dbe8f63c-7cac-47fb-b961-58c717e7a951\") " pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249853 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/14d2bb1f-8844-48cd-ba55-0782b444064e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249869 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d607e6e-db54-4e24-aafa-8fa7c16c949b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249885 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14d2bb1f-8844-48cd-ba55-0782b444064e-audit-dir\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.249901 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c59hf\" (UniqueName: \"kubernetes.io/projected/88d50f89-59b7-4772-8f49-7f7bfcc787ff-kube-api-access-c59hf\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.250171 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6d607e6e-db54-4e24-aafa-8fa7c16c949b-images\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.250234 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.250258 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-service-ca\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.253338 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14d2bb1f-8844-48cd-ba55-0782b444064e-audit-policies\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.254340 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a04cd83-dbd7-4076-858a-ff36748ed2d1-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.254892 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/88d50f89-59b7-4772-8f49-7f7bfcc787ff-etcd-serving-ca\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.257700 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/14d2bb1f-8844-48cd-ba55-0782b444064e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.258680 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c72540e8-b2dd-4377-8a17-717fcb2a05a2-serving-cert\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.258708 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.258749 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.259414 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14d2bb1f-8844-48cd-ba55-0782b444064e-audit-dir\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.259784 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wk2s"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.261080 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9r5mx"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.262398 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.267288 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2q455"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.270685 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.270906 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.272632 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.273022 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-serving-cert\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.274947 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/88d50f89-59b7-4772-8f49-7f7bfcc787ff-etcd-client\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.274988 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14d2bb1f-8844-48cd-ba55-0782b444064e-serving-cert\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.275183 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d607e6e-db54-4e24-aafa-8fa7c16c949b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.275406 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a04cd83-dbd7-4076-858a-ff36748ed2d1-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.275446 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/88d50f89-59b7-4772-8f49-7f7bfcc787ff-encryption-config\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.275485 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/021e5871-36e8-48d3-92de-e504a9429d9e-serving-cert\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.276550 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-4fdmd"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.277145 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4fdmd" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.278835 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/14d2bb1f-8844-48cd-ba55-0782b444064e-encryption-config\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.280519 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lxmpg"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.281543 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.281715 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.283113 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88d50f89-59b7-4772-8f49-7f7bfcc787ff-serving-cert\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.283179 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.285220 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.286685 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vgclk"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.287370 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.287560 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.294649 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/14d2bb1f-8844-48cd-ba55-0782b444064e-etcd-client\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.295480 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.304036 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-prnp7"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.306019 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.306170 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.307561 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pn9hw"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.314972 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-x6xql"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.315694 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f7lqb"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.315722 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.315734 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.315813 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.317366 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h5dwl"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.318563 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8ttx7"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.319346 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.320211 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.321371 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mmjn9"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.323548 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-b9fb4"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.326202 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.329944 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-n9zm5"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.330007 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.334373 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.334758 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fn6w5"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.336364 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.338537 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c54b8"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.342693 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.344254 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x6xql"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.346174 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.346183 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lxmpg"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.347239 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.348347 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.349421 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4fdmd"] Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.350906 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t5sd\" (UniqueName: \"kubernetes.io/projected/82bbeaf3-64ae-4676-a365-d75fdf225448-kube-api-access-6t5sd\") pod \"downloads-7954f5f757-9r5mx\" (UID: \"82bbeaf3-64ae-4676-a365-d75fdf225448\") " pod="openshift-console/downloads-7954f5f757-9r5mx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.350940 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/289056ee-b4e0-4eda-b826-5112ba8e54c0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtmvv\" (UID: \"289056ee-b4e0-4eda-b826-5112ba8e54c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.350958 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-serving-cert\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.350978 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/289056ee-b4e0-4eda-b826-5112ba8e54c0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtmvv\" (UID: \"289056ee-b4e0-4eda-b826-5112ba8e54c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.350993 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-drt6k\" (UID: \"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351012 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsh8m\" (UniqueName: \"kubernetes.io/projected/32d2100b-5750-4681-a3c7-5890f03107b7-kube-api-access-wsh8m\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351028 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1-config\") pod \"kube-controller-manager-operator-78b949d7b-drt6k\" (UID: \"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351044 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351060 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g45h5\" (UniqueName: \"kubernetes.io/projected/952902f2-2371-4a82-84e7-365978f206f6-kube-api-access-g45h5\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351076 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jjm6\" (UniqueName: \"kubernetes.io/projected/2fe0947a-c14b-498c-ba57-c5f8733ae76f-kube-api-access-8jjm6\") pod \"control-plane-machine-set-operator-78cbb6b69f-5js5f\" (UID: \"2fe0947a-c14b-498c-ba57-c5f8733ae76f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351125 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351168 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7lsm\" (UniqueName: \"kubernetes.io/projected/cdddbf25-7363-4abe-8363-90ed28ff079e-kube-api-access-r7lsm\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351197 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-drt6k\" (UID: \"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351225 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxkht\" (UniqueName: \"kubernetes.io/projected/1510c298-3548-442a-af69-72a237934ef4-kube-api-access-mxkht\") pod \"openshift-config-operator-7777fb866f-2q455\" (UID: \"1510c298-3548-442a-af69-72a237934ef4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351251 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3322e445-5258-46de-8fda-0c567244c1fa-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5v5vj\" (UID: \"3322e445-5258-46de-8fda-0c567244c1fa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351273 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54qwr\" (UniqueName: \"kubernetes.io/projected/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-kube-api-access-54qwr\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351317 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/952902f2-2371-4a82-84e7-365978f206f6-audit-dir\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351342 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2fe0947a-c14b-498c-ba57-c5f8733ae76f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5js5f\" (UID: \"2fe0947a-c14b-498c-ba57-c5f8733ae76f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351373 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18c77677-7e1e-4194-b117-74081b151270-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dsxgt\" (UID: \"18c77677-7e1e-4194-b117-74081b151270\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351420 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30854660-94ff-476c-a865-d326c3874db9-serving-cert\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351443 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx8vl\" (UniqueName: \"kubernetes.io/projected/8bf9078b-7040-4b98-86d5-1199c124ccb1-kube-api-access-xx8vl\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351467 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351490 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1510c298-3548-442a-af69-72a237934ef4-serving-cert\") pod \"openshift-config-operator-7777fb866f-2q455\" (UID: \"1510c298-3548-442a-af69-72a237934ef4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351514 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2da083b-f94f-4b8f-9657-be52a3ad66fa-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-h4kkh\" (UID: \"c2da083b-f94f-4b8f-9657-be52a3ad66fa\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351541 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-oauth-config\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351566 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351609 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/18c77677-7e1e-4194-b117-74081b151270-proxy-tls\") pod \"machine-config-controller-84d6567774-dsxgt\" (UID: \"18c77677-7e1e-4194-b117-74081b151270\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351632 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32d2100b-5750-4681-a3c7-5890f03107b7-service-ca-bundle\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351667 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggbsv\" (UniqueName: \"kubernetes.io/projected/205c1018-8396-442c-b265-42ee18193eba-kube-api-access-ggbsv\") pod \"cluster-samples-operator-665b6dd947-cdq5t\" (UID: \"205c1018-8396-442c-b265-42ee18193eba\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351689 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-auth-proxy-config\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351712 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351735 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e4ea004b-2e51-4b67-9238-b3fbe4575d72-srv-cert\") pod \"olm-operator-6b444d44fb-4x8qb\" (UID: \"e4ea004b-2e51-4b67-9238-b3fbe4575d72\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351765 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30854660-94ff-476c-a865-d326c3874db9-trusted-ca\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351788 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dbe8f63c-7cac-47fb-b961-58c717e7a951-metrics-tls\") pod \"dns-operator-744455d44c-pn9hw\" (UID: \"dbe8f63c-7cac-47fb-b961-58c717e7a951\") " pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351800 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1-config\") pod \"kube-controller-manager-operator-78b949d7b-drt6k\" (UID: \"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351809 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-service-ca\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351850 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351873 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cdddbf25-7363-4abe-8363-90ed28ff079e-images\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351894 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/32d2100b-5750-4681-a3c7-5890f03107b7-default-certificate\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351910 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-config\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351926 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351945 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r254\" (UniqueName: \"kubernetes.io/projected/e4ea004b-2e51-4b67-9238-b3fbe4575d72-kube-api-access-2r254\") pod \"olm-operator-6b444d44fb-4x8qb\" (UID: \"e4ea004b-2e51-4b67-9238-b3fbe4575d72\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351961 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2da083b-f94f-4b8f-9657-be52a3ad66fa-config\") pod \"kube-apiserver-operator-766d6c64bb-h4kkh\" (UID: \"c2da083b-f94f-4b8f-9657-be52a3ad66fa\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.351985 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cdddbf25-7363-4abe-8363-90ed28ff079e-proxy-tls\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352004 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-machine-approver-tls\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352019 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32d2100b-5750-4681-a3c7-5890f03107b7-metrics-certs\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352037 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/80398e72-5e3e-44ca-a188-8fd4db2ec02e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-prnp7\" (UID: \"80398e72-5e3e-44ca-a188-8fd4db2ec02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352067 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rdpj\" (UniqueName: \"kubernetes.io/projected/30854660-94ff-476c-a865-d326c3874db9-kube-api-access-8rdpj\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352097 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30854660-94ff-476c-a865-d326c3874db9-config\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352114 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352130 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9wpv\" (UniqueName: \"kubernetes.io/projected/ec8d8db8-33c4-433c-8d67-67546bd93235-kube-api-access-n9wpv\") pod \"package-server-manager-789f6589d5-r5v7h\" (UID: \"ec8d8db8-33c4-433c-8d67-67546bd93235\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352148 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-oauth-serving-cert\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352167 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qrhm\" (UniqueName: \"kubernetes.io/projected/289056ee-b4e0-4eda-b826-5112ba8e54c0-kube-api-access-4qrhm\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtmvv\" (UID: \"289056ee-b4e0-4eda-b826-5112ba8e54c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352184 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352199 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e4ea004b-2e51-4b67-9238-b3fbe4575d72-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4x8qb\" (UID: \"e4ea004b-2e51-4b67-9238-b3fbe4575d72\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352215 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-audit-policies\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352231 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5b9m\" (UniqueName: \"kubernetes.io/projected/18c77677-7e1e-4194-b117-74081b151270-kube-api-access-z5b9m\") pod \"machine-config-controller-84d6567774-dsxgt\" (UID: \"18c77677-7e1e-4194-b117-74081b151270\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352246 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2da083b-f94f-4b8f-9657-be52a3ad66fa-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-h4kkh\" (UID: \"c2da083b-f94f-4b8f-9657-be52a3ad66fa\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352304 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mrcs\" (UniqueName: \"kubernetes.io/projected/3322e445-5258-46de-8fda-0c567244c1fa-kube-api-access-7mrcs\") pod \"openshift-apiserver-operator-796bbdcf4f-5v5vj\" (UID: \"3322e445-5258-46de-8fda-0c567244c1fa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352322 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cdddbf25-7363-4abe-8363-90ed28ff079e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352338 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85lrl\" (UniqueName: \"kubernetes.io/projected/dbe8f63c-7cac-47fb-b961-58c717e7a951-kube-api-access-85lrl\") pod \"dns-operator-744455d44c-pn9hw\" (UID: \"dbe8f63c-7cac-47fb-b961-58c717e7a951\") " pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352354 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352371 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec8d8db8-33c4-433c-8d67-67546bd93235-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r5v7h\" (UID: \"ec8d8db8-33c4-433c-8d67-67546bd93235\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352390 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/205c1018-8396-442c-b265-42ee18193eba-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cdq5t\" (UID: \"205c1018-8396-442c-b265-42ee18193eba\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352406 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352422 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/32d2100b-5750-4681-a3c7-5890f03107b7-stats-auth\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352438 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-trusted-ca-bundle\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352454 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr4bc\" (UniqueName: \"kubernetes.io/projected/80398e72-5e3e-44ca-a188-8fd4db2ec02e-kube-api-access-lr4bc\") pod \"multus-admission-controller-857f4d67dd-prnp7\" (UID: \"80398e72-5e3e-44ca-a188-8fd4db2ec02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352473 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3322e445-5258-46de-8fda-0c567244c1fa-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5v5vj\" (UID: \"3322e445-5258-46de-8fda-0c567244c1fa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352490 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-config\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352506 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1510c298-3548-442a-af69-72a237934ef4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2q455\" (UID: \"1510c298-3548-442a-af69-72a237934ef4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352556 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/289056ee-b4e0-4eda-b826-5112ba8e54c0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtmvv\" (UID: \"289056ee-b4e0-4eda-b826-5112ba8e54c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352666 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-service-ca\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.352848 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1510c298-3548-442a-af69-72a237934ef4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2q455\" (UID: \"1510c298-3548-442a-af69-72a237934ef4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.354156 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-serving-cert\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.354170 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-oauth-serving-cert\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.355613 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/289056ee-b4e0-4eda-b826-5112ba8e54c0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtmvv\" (UID: \"289056ee-b4e0-4eda-b826-5112ba8e54c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.356187 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-audit-policies\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.356897 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.357332 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cdddbf25-7363-4abe-8363-90ed28ff079e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.357990 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.358488 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-trusted-ca-bundle\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.358683 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3322e445-5258-46de-8fda-0c567244c1fa-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5v5vj\" (UID: \"3322e445-5258-46de-8fda-0c567244c1fa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.358861 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.358970 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/952902f2-2371-4a82-84e7-365978f206f6-audit-dir\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.360023 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.360400 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-config\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.360402 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/205c1018-8396-442c-b265-42ee18193eba-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cdq5t\" (UID: \"205c1018-8396-442c-b265-42ee18193eba\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.360558 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cdddbf25-7363-4abe-8363-90ed28ff079e-images\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.360801 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1510c298-3548-442a-af69-72a237934ef4-serving-cert\") pod \"openshift-config-operator-7777fb866f-2q455\" (UID: \"1510c298-3548-442a-af69-72a237934ef4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.361157 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.361207 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-auth-proxy-config\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.361296 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-config\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.362006 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30854660-94ff-476c-a865-d326c3874db9-trusted-ca\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.362157 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.362228 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30854660-94ff-476c-a865-d326c3874db9-config\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.363268 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.363790 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.363811 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3322e445-5258-46de-8fda-0c567244c1fa-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5v5vj\" (UID: \"3322e445-5258-46de-8fda-0c567244c1fa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.363963 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-drt6k\" (UID: \"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.364023 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-oauth-config\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.365010 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dbe8f63c-7cac-47fb-b961-58c717e7a951-metrics-tls\") pod \"dns-operator-744455d44c-pn9hw\" (UID: \"dbe8f63c-7cac-47fb-b961-58c717e7a951\") " pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.365834 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.366674 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30854660-94ff-476c-a865-d326c3874db9-serving-cert\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.366944 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.367502 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.368816 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cdddbf25-7363-4abe-8363-90ed28ff079e-proxy-tls\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.368918 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.369143 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-machine-approver-tls\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.385781 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.391754 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2fe0947a-c14b-498c-ba57-c5f8733ae76f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5js5f\" (UID: \"2fe0947a-c14b-498c-ba57-c5f8733ae76f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.406328 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.445939 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.453798 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18c77677-7e1e-4194-b117-74081b151270-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dsxgt\" (UID: \"18c77677-7e1e-4194-b117-74081b151270\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.453857 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2da083b-f94f-4b8f-9657-be52a3ad66fa-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-h4kkh\" (UID: \"c2da083b-f94f-4b8f-9657-be52a3ad66fa\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454246 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/18c77677-7e1e-4194-b117-74081b151270-proxy-tls\") pod \"machine-config-controller-84d6567774-dsxgt\" (UID: \"18c77677-7e1e-4194-b117-74081b151270\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454280 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32d2100b-5750-4681-a3c7-5890f03107b7-service-ca-bundle\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454305 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e4ea004b-2e51-4b67-9238-b3fbe4575d72-srv-cert\") pod \"olm-operator-6b444d44fb-4x8qb\" (UID: \"e4ea004b-2e51-4b67-9238-b3fbe4575d72\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454363 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/32d2100b-5750-4681-a3c7-5890f03107b7-default-certificate\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454388 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r254\" (UniqueName: \"kubernetes.io/projected/e4ea004b-2e51-4b67-9238-b3fbe4575d72-kube-api-access-2r254\") pod \"olm-operator-6b444d44fb-4x8qb\" (UID: \"e4ea004b-2e51-4b67-9238-b3fbe4575d72\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454413 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2da083b-f94f-4b8f-9657-be52a3ad66fa-config\") pod \"kube-apiserver-operator-766d6c64bb-h4kkh\" (UID: \"c2da083b-f94f-4b8f-9657-be52a3ad66fa\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454436 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32d2100b-5750-4681-a3c7-5890f03107b7-metrics-certs\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454491 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/80398e72-5e3e-44ca-a188-8fd4db2ec02e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-prnp7\" (UID: \"80398e72-5e3e-44ca-a188-8fd4db2ec02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454548 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9wpv\" (UniqueName: \"kubernetes.io/projected/ec8d8db8-33c4-433c-8d67-67546bd93235-kube-api-access-n9wpv\") pod \"package-server-manager-789f6589d5-r5v7h\" (UID: \"ec8d8db8-33c4-433c-8d67-67546bd93235\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454613 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e4ea004b-2e51-4b67-9238-b3fbe4575d72-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4x8qb\" (UID: \"e4ea004b-2e51-4b67-9238-b3fbe4575d72\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454637 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5b9m\" (UniqueName: \"kubernetes.io/projected/18c77677-7e1e-4194-b117-74081b151270-kube-api-access-z5b9m\") pod \"machine-config-controller-84d6567774-dsxgt\" (UID: \"18c77677-7e1e-4194-b117-74081b151270\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454666 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2da083b-f94f-4b8f-9657-be52a3ad66fa-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-h4kkh\" (UID: \"c2da083b-f94f-4b8f-9657-be52a3ad66fa\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454715 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec8d8db8-33c4-433c-8d67-67546bd93235-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r5v7h\" (UID: \"ec8d8db8-33c4-433c-8d67-67546bd93235\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454742 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/32d2100b-5750-4681-a3c7-5890f03107b7-stats-auth\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454769 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr4bc\" (UniqueName: \"kubernetes.io/projected/80398e72-5e3e-44ca-a188-8fd4db2ec02e-kube-api-access-lr4bc\") pod \"multus-admission-controller-857f4d67dd-prnp7\" (UID: \"80398e72-5e3e-44ca-a188-8fd4db2ec02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454792 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18c77677-7e1e-4194-b117-74081b151270-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dsxgt\" (UID: \"18c77677-7e1e-4194-b117-74081b151270\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.454851 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsh8m\" (UniqueName: \"kubernetes.io/projected/32d2100b-5750-4681-a3c7-5890f03107b7-kube-api-access-wsh8m\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.465793 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.485900 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.505672 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.525522 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.546401 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.566522 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.586669 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.598135 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e4ea004b-2e51-4b67-9238-b3fbe4575d72-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4x8qb\" (UID: \"e4ea004b-2e51-4b67-9238-b3fbe4575d72\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.606983 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.626387 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.647756 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.666352 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.678726 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2da083b-f94f-4b8f-9657-be52a3ad66fa-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-h4kkh\" (UID: \"c2da083b-f94f-4b8f-9657-be52a3ad66fa\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.686211 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.706156 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.715385 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2da083b-f94f-4b8f-9657-be52a3ad66fa-config\") pod \"kube-apiserver-operator-766d6c64bb-h4kkh\" (UID: \"c2da083b-f94f-4b8f-9657-be52a3ad66fa\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.726756 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.737366 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/32d2100b-5750-4681-a3c7-5890f03107b7-default-certificate\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.746384 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.758124 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/32d2100b-5750-4681-a3c7-5890f03107b7-stats-auth\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.765922 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.777778 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32d2100b-5750-4681-a3c7-5890f03107b7-metrics-certs\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.785924 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.805613 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.815513 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32d2100b-5750-4681-a3c7-5890f03107b7-service-ca-bundle\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.826624 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.846723 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.866122 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.885901 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.897651 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/80398e72-5e3e-44ca-a188-8fd4db2ec02e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-prnp7\" (UID: \"80398e72-5e3e-44ca-a188-8fd4db2ec02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.905736 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.925583 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.945982 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.958422 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ec8d8db8-33c4-433c-8d67-67546bd93235-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r5v7h\" (UID: \"ec8d8db8-33c4-433c-8d67-67546bd93235\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.965497 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 04 04:36:27 crc kubenswrapper[4575]: I1004 04:36:27.993929 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.006853 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.026955 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.046029 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.065678 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.085481 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.105176 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.125826 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.137804 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e4ea004b-2e51-4b67-9238-b3fbe4575d72-srv-cert\") pod \"olm-operator-6b444d44fb-4x8qb\" (UID: \"e4ea004b-2e51-4b67-9238-b3fbe4575d72\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.146814 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.157874 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/18c77677-7e1e-4194-b117-74081b151270-proxy-tls\") pod \"machine-config-controller-84d6567774-dsxgt\" (UID: \"18c77677-7e1e-4194-b117-74081b151270\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.166394 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.186308 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.206471 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.224450 4575 request.go:700] Waited for 1.014108342s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.226469 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: E1004 04:36:28.239726 4575 configmap.go:193] Couldn't get configMap openshift-authentication-operator/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:36:28 crc kubenswrapper[4575]: E1004 04:36:28.239814 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-service-ca-bundle podName:4c57f905-7adb-4893-bb65-2f2f0adc8b5d nodeName:}" failed. No retries permitted until 2025-10-04 04:36:28.73979313 +0000 UTC m=+140.068351944 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-service-ca-bundle") pod "authentication-operator-69f744f599-mmjn9" (UID: "4c57f905-7adb-4893-bb65-2f2f0adc8b5d") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.246698 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.266772 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.286130 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.306432 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.326021 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.346680 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.366295 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.386902 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.406159 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.426492 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.446506 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.466090 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.485371 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.506405 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.525784 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.545335 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.566529 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.586925 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.614953 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.625760 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.645843 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.665812 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.686705 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.705946 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.726001 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.746410 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.766388 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.771693 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-service-ca-bundle\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.785626 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.820708 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hknmw\" (UniqueName: \"kubernetes.io/projected/4a04cd83-dbd7-4076-858a-ff36748ed2d1-kube-api-access-hknmw\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.840701 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls796\" (UniqueName: \"kubernetes.io/projected/021e5871-36e8-48d3-92de-e504a9429d9e-kube-api-access-ls796\") pod \"route-controller-manager-6576b87f9c-t2mds\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.860525 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cdkp\" (UniqueName: \"kubernetes.io/projected/c72540e8-b2dd-4377-8a17-717fcb2a05a2-kube-api-access-8cdkp\") pod \"controller-manager-879f6c89f-flrql\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.882887 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hgrg\" (UniqueName: \"kubernetes.io/projected/14d2bb1f-8844-48cd-ba55-0782b444064e-kube-api-access-6hgrg\") pod \"apiserver-7bbb656c7d-jlb6g\" (UID: \"14d2bb1f-8844-48cd-ba55-0782b444064e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.903976 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a04cd83-dbd7-4076-858a-ff36748ed2d1-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rmkmw\" (UID: \"4a04cd83-dbd7-4076-858a-ff36748ed2d1\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.904139 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.925989 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c59hf\" (UniqueName: \"kubernetes.io/projected/88d50f89-59b7-4772-8f49-7f7bfcc787ff-kube-api-access-c59hf\") pod \"apiserver-76f77b778f-q9ssb\" (UID: \"88d50f89-59b7-4772-8f49-7f7bfcc787ff\") " pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.933686 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.942723 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhtvr\" (UniqueName: \"kubernetes.io/projected/6d607e6e-db54-4e24-aafa-8fa7c16c949b-kube-api-access-qhtvr\") pod \"machine-api-operator-5694c8668f-rdr7j\" (UID: \"6d607e6e-db54-4e24-aafa-8fa7c16c949b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.954033 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.985898 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 04 04:36:28 crc kubenswrapper[4575]: I1004 04:36:28.993298 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.007222 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.029499 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.047408 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.066569 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.086360 4575 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.105915 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.126721 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.145163 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.146597 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.152427 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds"] Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.165778 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 04 04:36:29 crc kubenswrapper[4575]: W1004 04:36:29.169477 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod021e5871_36e8_48d3_92de_e504a9429d9e.slice/crio-e6695130a87c43b98c1b1e876d378e44592156ecc618cb09241abcdd7b671a06 WatchSource:0}: Error finding container e6695130a87c43b98c1b1e876d378e44592156ecc618cb09241abcdd7b671a06: Status 404 returned error can't find the container with id e6695130a87c43b98c1b1e876d378e44592156ecc618cb09241abcdd7b671a06 Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.174125 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.176752 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flrql"] Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.187787 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.198281 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rdr7j"] Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.207160 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.226835 4575 request.go:700] Waited for 1.907115115s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-dockercfg-qx5rd&limit=500&resourceVersion=0 Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.231904 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.244394 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw"] Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.271142 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t5sd\" (UniqueName: \"kubernetes.io/projected/82bbeaf3-64ae-4676-a365-d75fdf225448-kube-api-access-6t5sd\") pod \"downloads-7954f5f757-9r5mx\" (UID: \"82bbeaf3-64ae-4676-a365-d75fdf225448\") " pod="openshift-console/downloads-7954f5f757-9r5mx" Oct 04 04:36:29 crc kubenswrapper[4575]: W1004 04:36:29.290347 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a04cd83_dbd7_4076_858a_ff36748ed2d1.slice/crio-b20ed7e0c7c72f77a6776cc2ba8535cc357b6171111c265485609eaa79616471 WatchSource:0}: Error finding container b20ed7e0c7c72f77a6776cc2ba8535cc357b6171111c265485609eaa79616471: Status 404 returned error can't find the container with id b20ed7e0c7c72f77a6776cc2ba8535cc357b6171111c265485609eaa79616471 Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.294923 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jjm6\" (UniqueName: \"kubernetes.io/projected/2fe0947a-c14b-498c-ba57-c5f8733ae76f-kube-api-access-8jjm6\") pod \"control-plane-machine-set-operator-78cbb6b69f-5js5f\" (UID: \"2fe0947a-c14b-498c-ba57-c5f8733ae76f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.319437 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-drt6k\" (UID: \"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.325191 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qrhm\" (UniqueName: \"kubernetes.io/projected/289056ee-b4e0-4eda-b826-5112ba8e54c0-kube-api-access-4qrhm\") pod \"openshift-controller-manager-operator-756b6f6bc6-gtmvv\" (UID: \"289056ee-b4e0-4eda-b826-5112ba8e54c0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.343971 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mrcs\" (UniqueName: \"kubernetes.io/projected/3322e445-5258-46de-8fda-0c567244c1fa-kube-api-access-7mrcs\") pod \"openshift-apiserver-operator-796bbdcf4f-5v5vj\" (UID: \"3322e445-5258-46de-8fda-0c567244c1fa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.368438 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7lsm\" (UniqueName: \"kubernetes.io/projected/cdddbf25-7363-4abe-8363-90ed28ff079e-kube-api-access-r7lsm\") pod \"machine-config-operator-74547568cd-qmhlx\" (UID: \"cdddbf25-7363-4abe-8363-90ed28ff079e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.392550 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85lrl\" (UniqueName: \"kubernetes.io/projected/dbe8f63c-7cac-47fb-b961-58c717e7a951-kube-api-access-85lrl\") pod \"dns-operator-744455d44c-pn9hw\" (UID: \"dbe8f63c-7cac-47fb-b961-58c717e7a951\") " pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.407285 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.409215 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxkht\" (UniqueName: \"kubernetes.io/projected/1510c298-3548-442a-af69-72a237934ef4-kube-api-access-mxkht\") pod \"openshift-config-operator-7777fb866f-2q455\" (UID: \"1510c298-3548-442a-af69-72a237934ef4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.415516 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g"] Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.433157 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9r5mx" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.440977 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g45h5\" (UniqueName: \"kubernetes.io/projected/952902f2-2371-4a82-84e7-365978f206f6-kube-api-access-g45h5\") pod \"oauth-openshift-558db77b4-2wk2s\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.445193 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54qwr\" (UniqueName: \"kubernetes.io/projected/7cf2b5e5-0623-438d-bee5-f2403d7f6e43-kube-api-access-54qwr\") pod \"machine-approver-56656f9798-q6xg8\" (UID: \"7cf2b5e5-0623-438d-bee5-f2403d7f6e43\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.447267 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.447631 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q9ssb"] Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.455353 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.460492 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx8vl\" (UniqueName: \"kubernetes.io/projected/8bf9078b-7040-4b98-86d5-1199c124ccb1-kube-api-access-xx8vl\") pod \"console-f9d7485db-vgclk\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:29 crc kubenswrapper[4575]: W1004 04:36:29.464677 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88d50f89_59b7_4772_8f49_7f7bfcc787ff.slice/crio-fc50214104bd320040f4afd0c423360c4d4d6945ab5062eb81ad0974fa7f6711 WatchSource:0}: Error finding container fc50214104bd320040f4afd0c423360c4d4d6945ab5062eb81ad0974fa7f6711: Status 404 returned error can't find the container with id fc50214104bd320040f4afd0c423360c4d4d6945ab5062eb81ad0974fa7f6711 Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.487140 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggbsv\" (UniqueName: \"kubernetes.io/projected/205c1018-8396-442c-b265-42ee18193eba-kube-api-access-ggbsv\") pod \"cluster-samples-operator-665b6dd947-cdq5t\" (UID: \"205c1018-8396-442c-b265-42ee18193eba\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.487543 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.492795 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.501053 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.512102 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rdpj\" (UniqueName: \"kubernetes.io/projected/30854660-94ff-476c-a865-d326c3874db9-kube-api-access-8rdpj\") pod \"console-operator-58897d9998-f7lqb\" (UID: \"30854660-94ff-476c-a865-d326c3874db9\") " pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.540736 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r254\" (UniqueName: \"kubernetes.io/projected/e4ea004b-2e51-4b67-9238-b3fbe4575d72-kube-api-access-2r254\") pod \"olm-operator-6b444d44fb-4x8qb\" (UID: \"e4ea004b-2e51-4b67-9238-b3fbe4575d72\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.573342 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.577955 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9wpv\" (UniqueName: \"kubernetes.io/projected/ec8d8db8-33c4-433c-8d67-67546bd93235-kube-api-access-n9wpv\") pod \"package-server-manager-789f6589d5-r5v7h\" (UID: \"ec8d8db8-33c4-433c-8d67-67546bd93235\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.602249 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5b9m\" (UniqueName: \"kubernetes.io/projected/18c77677-7e1e-4194-b117-74081b151270-kube-api-access-z5b9m\") pod \"machine-config-controller-84d6567774-dsxgt\" (UID: \"18c77677-7e1e-4194-b117-74081b151270\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.607424 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.608866 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2da083b-f94f-4b8f-9657-be52a3ad66fa-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-h4kkh\" (UID: \"c2da083b-f94f-4b8f-9657-be52a3ad66fa\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.625046 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr4bc\" (UniqueName: \"kubernetes.io/projected/80398e72-5e3e-44ca-a188-8fd4db2ec02e-kube-api-access-lr4bc\") pod \"multus-admission-controller-857f4d67dd-prnp7\" (UID: \"80398e72-5e3e-44ca-a188-8fd4db2ec02e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.625705 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.632095 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.641957 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.648954 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.649972 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsh8m\" (UniqueName: \"kubernetes.io/projected/32d2100b-5750-4681-a3c7-5890f03107b7-kube-api-access-wsh8m\") pod \"router-default-5444994796-bd5sj\" (UID: \"32d2100b-5750-4681-a3c7-5890f03107b7\") " pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.667398 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.686217 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.689810 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sswjt\" (UniqueName: \"kubernetes.io/projected/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-kube-api-access-sswjt\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.694639 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c57f905-7adb-4893-bb65-2f2f0adc8b5d-service-ca-bundle\") pod \"authentication-operator-69f744f599-mmjn9\" (UID: \"4c57f905-7adb-4893-bb65-2f2f0adc8b5d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.716340 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.733320 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj"] Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.738310 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" Oct 04 04:36:29 crc kubenswrapper[4575]: W1004 04:36:29.751259 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cf2b5e5_0623_438d_bee5_f2403d7f6e43.slice/crio-ae8c5d9e660ce5ca38c98bfe11e849c36847e4ca4273775486cce8641f1f388f WatchSource:0}: Error finding container ae8c5d9e660ce5ca38c98bfe11e849c36847e4ca4273775486cce8641f1f388f: Status 404 returned error can't find the container with id ae8c5d9e660ce5ca38c98bfe11e849c36847e4ca4273775486cce8641f1f388f Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.796854 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf4da93-109f-4201-8fde-205d3606733f-config\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.796900 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0a86ffc6-40bc-4d9b-b23f-37bac308a563-signing-cabundle\") pod \"service-ca-9c57cc56f-n9zm5\" (UID: \"0a86ffc6-40bc-4d9b-b23f-37bac308a563\") " pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.796959 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmlpv\" (UniqueName: \"kubernetes.io/projected/c48ef73d-9496-48ae-98a5-7a3fbfe60187-kube-api-access-xmlpv\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797001 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4caea03c-cd82-4798-b8b1-7007cc09079a-secret-volume\") pod \"collect-profiles-29325870-d7k6g\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797021 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c48ef73d-9496-48ae-98a5-7a3fbfe60187-webhook-cert\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797043 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt6h9\" (UniqueName: \"kubernetes.io/projected/4caea03c-cd82-4798-b8b1-7007cc09079a-kube-api-access-mt6h9\") pod \"collect-profiles-29325870-d7k6g\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797105 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c48ef73d-9496-48ae-98a5-7a3fbfe60187-tmpfs\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797128 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f6eaec-4e9a-4c19-b95f-cf103db54fdc-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x46h8\" (UID: \"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797183 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-tls\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797208 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xxs96\" (UID: \"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797258 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797281 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fn6w5\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797347 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797434 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0b2bce21-b555-4085-9d07-a1da0661bab6-profile-collector-cert\") pod \"catalog-operator-68c6474976-b9zwp\" (UID: \"0b2bce21-b555-4085-9d07-a1da0661bab6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797471 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4caea03c-cd82-4798-b8b1-7007cc09079a-config-volume\") pod \"collect-profiles-29325870-d7k6g\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797493 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr9bw\" (UniqueName: \"kubernetes.io/projected/0b2bce21-b555-4085-9d07-a1da0661bab6-kube-api-access-lr9bw\") pod \"catalog-operator-68c6474976-b9zwp\" (UID: \"0b2bce21-b555-4085-9d07-a1da0661bab6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797529 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797567 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0a86ffc6-40bc-4d9b-b23f-37bac308a563-signing-key\") pod \"service-ca-9c57cc56f-n9zm5\" (UID: \"0a86ffc6-40bc-4d9b-b23f-37bac308a563\") " pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797655 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8bf4da93-109f-4201-8fde-205d3606733f-etcd-service-ca\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797721 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e12a025a-24e2-4860-847c-4c2ec660015c-metrics-tls\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797747 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9zv2\" (UniqueName: \"kubernetes.io/projected/844cd2d0-93d0-45d9-b342-b86df49548d8-kube-api-access-m9zv2\") pod \"marketplace-operator-79b997595-fn6w5\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797767 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xxs96\" (UID: \"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797804 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/886b072f-91e0-42a4-ac24-6ce0b25956a9-serving-cert\") pod \"service-ca-operator-777779d784-c54b8\" (UID: \"886b072f-91e0-42a4-ac24-6ce0b25956a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797867 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c48ef73d-9496-48ae-98a5-7a3fbfe60187-apiservice-cert\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797903 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8bf4da93-109f-4201-8fde-205d3606733f-etcd-client\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.797972 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0b2bce21-b555-4085-9d07-a1da0661bab6-srv-cert\") pod \"catalog-operator-68c6474976-b9zwp\" (UID: \"0b2bce21-b555-4085-9d07-a1da0661bab6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798059 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzqnk\" (UniqueName: \"kubernetes.io/projected/0a86ffc6-40bc-4d9b-b23f-37bac308a563-kube-api-access-fzqnk\") pod \"service-ca-9c57cc56f-n9zm5\" (UID: \"0a86ffc6-40bc-4d9b-b23f-37bac308a563\") " pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798110 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8bf4da93-109f-4201-8fde-205d3606733f-etcd-ca\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798160 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xxs96\" (UID: \"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798209 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkj4k\" (UniqueName: \"kubernetes.io/projected/8bf4da93-109f-4201-8fde-205d3606733f-kube-api-access-dkj4k\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798232 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fn6w5\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798271 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9f6eaec-4e9a-4c19-b95f-cf103db54fdc-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x46h8\" (UID: \"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798291 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-trusted-ca\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798320 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-bound-sa-token\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798356 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8bf4da93-109f-4201-8fde-205d3606733f-serving-cert\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798376 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e12a025a-24e2-4860-847c-4c2ec660015c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798426 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886b072f-91e0-42a4-ac24-6ce0b25956a9-config\") pod \"service-ca-operator-777779d784-c54b8\" (UID: \"886b072f-91e0-42a4-ac24-6ce0b25956a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798470 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-certificates\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798509 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zqm8\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-kube-api-access-6zqm8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798538 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e12a025a-24e2-4860-847c-4c2ec660015c-trusted-ca\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798576 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5xgx\" (UniqueName: \"kubernetes.io/projected/d9f6eaec-4e9a-4c19-b95f-cf103db54fdc-kube-api-access-p5xgx\") pod \"kube-storage-version-migrator-operator-b67b599dd-x46h8\" (UID: \"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798632 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2plg\" (UniqueName: \"kubernetes.io/projected/886b072f-91e0-42a4-ac24-6ce0b25956a9-kube-api-access-f2plg\") pod \"service-ca-operator-777779d784-c54b8\" (UID: \"886b072f-91e0-42a4-ac24-6ce0b25956a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798706 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldppk\" (UniqueName: \"kubernetes.io/projected/e12a025a-24e2-4860-847c-4c2ec660015c-kube-api-access-ldppk\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.798766 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlf98\" (UniqueName: \"kubernetes.io/projected/800a54e1-61b2-4e0a-a9a4-423456d0fe78-kube-api-access-tlf98\") pod \"migrator-59844c95c7-8rt5h\" (UID: \"800a54e1-61b2-4e0a-a9a4-423456d0fe78\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h" Oct 04 04:36:29 crc kubenswrapper[4575]: E1004 04:36:29.805378 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:30.305361965 +0000 UTC m=+141.633920789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.819302 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.828228 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.842359 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k"] Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.842664 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.864867 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.868873 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.881418 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.899712 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.899968 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkj4k\" (UniqueName: \"kubernetes.io/projected/8bf4da93-109f-4201-8fde-205d3606733f-kube-api-access-dkj4k\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900003 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fn6w5\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900043 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9f6eaec-4e9a-4c19-b95f-cf103db54fdc-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x46h8\" (UID: \"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900066 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-trusted-ca\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900111 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-bound-sa-token\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900134 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8bf4da93-109f-4201-8fde-205d3606733f-serving-cert\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900158 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e12a025a-24e2-4860-847c-4c2ec660015c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900188 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/667c43a4-aeb4-46b4-996a-e0695695dbbd-node-bootstrap-token\") pod \"machine-config-server-8ttx7\" (UID: \"667c43a4-aeb4-46b4-996a-e0695695dbbd\") " pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900257 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-certificates\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900283 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886b072f-91e0-42a4-ac24-6ce0b25956a9-config\") pod \"service-ca-operator-777779d784-c54b8\" (UID: \"886b072f-91e0-42a4-ac24-6ce0b25956a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900306 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcddd\" (UniqueName: \"kubernetes.io/projected/eb94e3f3-e813-4122-8718-26d6e005c8f7-kube-api-access-tcddd\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900343 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zqm8\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-kube-api-access-6zqm8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900373 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86c9t\" (UniqueName: \"kubernetes.io/projected/638083de-01ec-4816-a985-db72a26ccce1-kube-api-access-86c9t\") pod \"dns-default-x6xql\" (UID: \"638083de-01ec-4816-a985-db72a26ccce1\") " pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900398 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-csi-data-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900424 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-plugins-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900456 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2plg\" (UniqueName: \"kubernetes.io/projected/886b072f-91e0-42a4-ac24-6ce0b25956a9-kube-api-access-f2plg\") pod \"service-ca-operator-777779d784-c54b8\" (UID: \"886b072f-91e0-42a4-ac24-6ce0b25956a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900482 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e12a025a-24e2-4860-847c-4c2ec660015c-trusted-ca\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900517 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5xgx\" (UniqueName: \"kubernetes.io/projected/d9f6eaec-4e9a-4c19-b95f-cf103db54fdc-kube-api-access-p5xgx\") pod \"kube-storage-version-migrator-operator-b67b599dd-x46h8\" (UID: \"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900559 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/616523ee-ffd4-454a-a527-1ef2838edaef-cert\") pod \"ingress-canary-4fdmd\" (UID: \"616523ee-ffd4-454a-a527-1ef2838edaef\") " pod="openshift-ingress-canary/ingress-canary-4fdmd" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900619 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldppk\" (UniqueName: \"kubernetes.io/projected/e12a025a-24e2-4860-847c-4c2ec660015c-kube-api-access-ldppk\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900673 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlf98\" (UniqueName: \"kubernetes.io/projected/800a54e1-61b2-4e0a-a9a4-423456d0fe78-kube-api-access-tlf98\") pod \"migrator-59844c95c7-8rt5h\" (UID: \"800a54e1-61b2-4e0a-a9a4-423456d0fe78\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900700 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-mountpoint-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900725 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0a86ffc6-40bc-4d9b-b23f-37bac308a563-signing-cabundle\") pod \"service-ca-9c57cc56f-n9zm5\" (UID: \"0a86ffc6-40bc-4d9b-b23f-37bac308a563\") " pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900761 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf4da93-109f-4201-8fde-205d3606733f-config\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900816 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4caea03c-cd82-4798-b8b1-7007cc09079a-secret-volume\") pod \"collect-profiles-29325870-d7k6g\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900841 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmlpv\" (UniqueName: \"kubernetes.io/projected/c48ef73d-9496-48ae-98a5-7a3fbfe60187-kube-api-access-xmlpv\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900879 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c48ef73d-9496-48ae-98a5-7a3fbfe60187-webhook-cert\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900903 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt6h9\" (UniqueName: \"kubernetes.io/projected/4caea03c-cd82-4798-b8b1-7007cc09079a-kube-api-access-mt6h9\") pod \"collect-profiles-29325870-d7k6g\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900965 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c48ef73d-9496-48ae-98a5-7a3fbfe60187-tmpfs\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.900990 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f6eaec-4e9a-4c19-b95f-cf103db54fdc-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x46h8\" (UID: \"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901029 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-tls\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901052 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901076 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xxs96\" (UID: \"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901116 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fn6w5\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901143 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-registration-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901187 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901239 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4mn7\" (UniqueName: \"kubernetes.io/projected/616523ee-ffd4-454a-a527-1ef2838edaef-kube-api-access-w4mn7\") pod \"ingress-canary-4fdmd\" (UID: \"616523ee-ffd4-454a-a527-1ef2838edaef\") " pod="openshift-ingress-canary/ingress-canary-4fdmd" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901292 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0b2bce21-b555-4085-9d07-a1da0661bab6-profile-collector-cert\") pod \"catalog-operator-68c6474976-b9zwp\" (UID: \"0b2bce21-b555-4085-9d07-a1da0661bab6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901315 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4caea03c-cd82-4798-b8b1-7007cc09079a-config-volume\") pod \"collect-profiles-29325870-d7k6g\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901361 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr9bw\" (UniqueName: \"kubernetes.io/projected/0b2bce21-b555-4085-9d07-a1da0661bab6-kube-api-access-lr9bw\") pod \"catalog-operator-68c6474976-b9zwp\" (UID: \"0b2bce21-b555-4085-9d07-a1da0661bab6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901466 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0a86ffc6-40bc-4d9b-b23f-37bac308a563-signing-key\") pod \"service-ca-9c57cc56f-n9zm5\" (UID: \"0a86ffc6-40bc-4d9b-b23f-37bac308a563\") " pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901522 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8bf4da93-109f-4201-8fde-205d3606733f-etcd-service-ca\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901551 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e12a025a-24e2-4860-847c-4c2ec660015c-metrics-tls\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901703 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/638083de-01ec-4816-a985-db72a26ccce1-config-volume\") pod \"dns-default-x6xql\" (UID: \"638083de-01ec-4816-a985-db72a26ccce1\") " pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901741 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9zv2\" (UniqueName: \"kubernetes.io/projected/844cd2d0-93d0-45d9-b342-b86df49548d8-kube-api-access-m9zv2\") pod \"marketplace-operator-79b997595-fn6w5\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901771 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xxs96\" (UID: \"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901814 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c48ef73d-9496-48ae-98a5-7a3fbfe60187-apiservice-cert\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901840 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/886b072f-91e0-42a4-ac24-6ce0b25956a9-serving-cert\") pod \"service-ca-operator-777779d784-c54b8\" (UID: \"886b072f-91e0-42a4-ac24-6ce0b25956a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901882 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8bf4da93-109f-4201-8fde-205d3606733f-etcd-client\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901906 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnns4\" (UniqueName: \"kubernetes.io/projected/667c43a4-aeb4-46b4-996a-e0695695dbbd-kube-api-access-mnns4\") pod \"machine-config-server-8ttx7\" (UID: \"667c43a4-aeb4-46b4-996a-e0695695dbbd\") " pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901934 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0b2bce21-b555-4085-9d07-a1da0661bab6-srv-cert\") pod \"catalog-operator-68c6474976-b9zwp\" (UID: \"0b2bce21-b555-4085-9d07-a1da0661bab6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.901959 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/638083de-01ec-4816-a985-db72a26ccce1-metrics-tls\") pod \"dns-default-x6xql\" (UID: \"638083de-01ec-4816-a985-db72a26ccce1\") " pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.902033 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzqnk\" (UniqueName: \"kubernetes.io/projected/0a86ffc6-40bc-4d9b-b23f-37bac308a563-kube-api-access-fzqnk\") pod \"service-ca-9c57cc56f-n9zm5\" (UID: \"0a86ffc6-40bc-4d9b-b23f-37bac308a563\") " pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.902057 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-socket-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.902099 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8bf4da93-109f-4201-8fde-205d3606733f-etcd-ca\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.902167 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/667c43a4-aeb4-46b4-996a-e0695695dbbd-certs\") pod \"machine-config-server-8ttx7\" (UID: \"667c43a4-aeb4-46b4-996a-e0695695dbbd\") " pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.902219 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xxs96\" (UID: \"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.905189 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: E1004 04:36:29.905319 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:30.405296407 +0000 UTC m=+141.733855231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.915421 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xxs96\" (UID: \"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.915455 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e12a025a-24e2-4860-847c-4c2ec660015c-trusted-ca\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.916334 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9f6eaec-4e9a-4c19-b95f-cf103db54fdc-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x46h8\" (UID: \"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.920398 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8bf4da93-109f-4201-8fde-205d3606733f-etcd-ca\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.929318 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fn6w5\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.924632 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e12a025a-24e2-4860-847c-4c2ec660015c-metrics-tls\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.933364 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c48ef73d-9496-48ae-98a5-7a3fbfe60187-apiservice-cert\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.938325 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f6eaec-4e9a-4c19-b95f-cf103db54fdc-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x46h8\" (UID: \"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.940319 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fn6w5\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.942036 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8bf4da93-109f-4201-8fde-205d3606733f-etcd-service-ca\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.948311 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4caea03c-cd82-4798-b8b1-7007cc09079a-secret-volume\") pod \"collect-profiles-29325870-d7k6g\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.951229 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c48ef73d-9496-48ae-98a5-7a3fbfe60187-webhook-cert\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.957019 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-trusted-ca\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.958652 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c48ef73d-9496-48ae-98a5-7a3fbfe60187-tmpfs\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.978121 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-tls\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.979424 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0a86ffc6-40bc-4d9b-b23f-37bac308a563-signing-cabundle\") pod \"service-ca-9c57cc56f-n9zm5\" (UID: \"0a86ffc6-40bc-4d9b-b23f-37bac308a563\") " pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.984493 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf4da93-109f-4201-8fde-205d3606733f-config\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.988202 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-certificates\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.989238 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.989514 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886b072f-91e0-42a4-ac24-6ce0b25956a9-config\") pod \"service-ca-operator-777779d784-c54b8\" (UID: \"886b072f-91e0-42a4-ac24-6ce0b25956a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.989737 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8bf4da93-109f-4201-8fde-205d3606733f-serving-cert\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.990568 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4caea03c-cd82-4798-b8b1-7007cc09079a-config-volume\") pod \"collect-profiles-29325870-d7k6g\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.994510 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8bf4da93-109f-4201-8fde-205d3606733f-etcd-client\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.995760 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xxs96\" (UID: \"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.998506 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" event={"ID":"c72540e8-b2dd-4377-8a17-717fcb2a05a2","Type":"ContainerStarted","Data":"a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f"} Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.998631 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" event={"ID":"c72540e8-b2dd-4377-8a17-717fcb2a05a2","Type":"ContainerStarted","Data":"df1f9a2773e11e042ffae27840b39f30c13fa3450c2c1fdb0fa007afd5fe06a1"} Oct 04 04:36:29 crc kubenswrapper[4575]: I1004 04:36:29.998662 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.001396 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" event={"ID":"88d50f89-59b7-4772-8f49-7f7bfcc787ff","Type":"ContainerStarted","Data":"fc50214104bd320040f4afd0c423360c4d4d6945ab5062eb81ad0974fa7f6711"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.003423 4575 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-flrql container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.003487 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" podUID="c72540e8-b2dd-4377-8a17-717fcb2a05a2" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.009700 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr9bw\" (UniqueName: \"kubernetes.io/projected/0b2bce21-b555-4085-9d07-a1da0661bab6-kube-api-access-lr9bw\") pod \"catalog-operator-68c6474976-b9zwp\" (UID: \"0b2bce21-b555-4085-9d07-a1da0661bab6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.010138 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-mountpoint-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.010367 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-registration-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.010829 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4mn7\" (UniqueName: \"kubernetes.io/projected/616523ee-ffd4-454a-a527-1ef2838edaef-kube-api-access-w4mn7\") pod \"ingress-canary-4fdmd\" (UID: \"616523ee-ffd4-454a-a527-1ef2838edaef\") " pod="openshift-ingress-canary/ingress-canary-4fdmd" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.017238 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.012710 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-registration-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.015131 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" event={"ID":"14d2bb1f-8844-48cd-ba55-0782b444064e","Type":"ContainerStarted","Data":"3fb682c531122a3bfdaa51614e941584af5b0c6a02724a62c9ba389d0b878a6d"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.030496 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" event={"ID":"6d607e6e-db54-4e24-aafa-8fa7c16c949b","Type":"ContainerStarted","Data":"87c946f2d494d3481d030330f6b966be656ad30b19d3b1eeb3a0da7230870b1f"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.030538 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" event={"ID":"6d607e6e-db54-4e24-aafa-8fa7c16c949b","Type":"ContainerStarted","Data":"8ab88e2d4a8cab418ca39ad640204c8727852b50620a866adb08f0a18173c3f1"} Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.019970 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:30.519940556 +0000 UTC m=+141.848499360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.030760 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/638083de-01ec-4816-a985-db72a26ccce1-config-volume\") pod \"dns-default-x6xql\" (UID: \"638083de-01ec-4816-a985-db72a26ccce1\") " pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.030851 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnns4\" (UniqueName: \"kubernetes.io/projected/667c43a4-aeb4-46b4-996a-e0695695dbbd-kube-api-access-mnns4\") pod \"machine-config-server-8ttx7\" (UID: \"667c43a4-aeb4-46b4-996a-e0695695dbbd\") " pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.030937 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/638083de-01ec-4816-a985-db72a26ccce1-metrics-tls\") pod \"dns-default-x6xql\" (UID: \"638083de-01ec-4816-a985-db72a26ccce1\") " pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.031079 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-socket-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.031131 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/667c43a4-aeb4-46b4-996a-e0695695dbbd-certs\") pod \"machine-config-server-8ttx7\" (UID: \"667c43a4-aeb4-46b4-996a-e0695695dbbd\") " pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.031505 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/667c43a4-aeb4-46b4-996a-e0695695dbbd-node-bootstrap-token\") pod \"machine-config-server-8ttx7\" (UID: \"667c43a4-aeb4-46b4-996a-e0695695dbbd\") " pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.031540 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcddd\" (UniqueName: \"kubernetes.io/projected/eb94e3f3-e813-4122-8718-26d6e005c8f7-kube-api-access-tcddd\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.031608 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86c9t\" (UniqueName: \"kubernetes.io/projected/638083de-01ec-4816-a985-db72a26ccce1-kube-api-access-86c9t\") pod \"dns-default-x6xql\" (UID: \"638083de-01ec-4816-a985-db72a26ccce1\") " pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.031772 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-csi-data-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.031815 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-plugins-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.010745 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-mountpoint-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.032323 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/616523ee-ffd4-454a-a527-1ef2838edaef-cert\") pod \"ingress-canary-4fdmd\" (UID: \"616523ee-ffd4-454a-a527-1ef2838edaef\") " pod="openshift-ingress-canary/ingress-canary-4fdmd" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.033129 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2plg\" (UniqueName: \"kubernetes.io/projected/886b072f-91e0-42a4-ac24-6ce0b25956a9-kube-api-access-f2plg\") pod \"service-ca-operator-777779d784-c54b8\" (UID: \"886b072f-91e0-42a4-ac24-6ce0b25956a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.033542 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-socket-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.033696 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-csi-data-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.034211 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/638083de-01ec-4816-a985-db72a26ccce1-config-volume\") pod \"dns-default-x6xql\" (UID: \"638083de-01ec-4816-a985-db72a26ccce1\") " pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.038406 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkj4k\" (UniqueName: \"kubernetes.io/projected/8bf4da93-109f-4201-8fde-205d3606733f-kube-api-access-dkj4k\") pod \"etcd-operator-b45778765-b9fb4\" (UID: \"8bf4da93-109f-4201-8fde-205d3606733f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.038949 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0a86ffc6-40bc-4d9b-b23f-37bac308a563-signing-key\") pod \"service-ca-9c57cc56f-n9zm5\" (UID: \"0a86ffc6-40bc-4d9b-b23f-37bac308a563\") " pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.040051 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/eb94e3f3-e813-4122-8718-26d6e005c8f7-plugins-dir\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.041719 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" event={"ID":"3322e445-5258-46de-8fda-0c567244c1fa","Type":"ContainerStarted","Data":"559aa9e01abbe098a29179c825988eee7bc03890d520f5fce9f90ff22ab90642"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.042005 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/886b072f-91e0-42a4-ac24-6ce0b25956a9-serving-cert\") pod \"service-ca-operator-777779d784-c54b8\" (UID: \"886b072f-91e0-42a4-ac24-6ce0b25956a9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.043151 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0b2bce21-b555-4085-9d07-a1da0661bab6-srv-cert\") pod \"catalog-operator-68c6474976-b9zwp\" (UID: \"0b2bce21-b555-4085-9d07-a1da0661bab6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.051346 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/616523ee-ffd4-454a-a527-1ef2838edaef-cert\") pod \"ingress-canary-4fdmd\" (UID: \"616523ee-ffd4-454a-a527-1ef2838edaef\") " pod="openshift-ingress-canary/ingress-canary-4fdmd" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.051894 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9zv2\" (UniqueName: \"kubernetes.io/projected/844cd2d0-93d0-45d9-b342-b86df49548d8-kube-api-access-m9zv2\") pod \"marketplace-operator-79b997595-fn6w5\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.052277 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" event={"ID":"4a04cd83-dbd7-4076-858a-ff36748ed2d1","Type":"ContainerStarted","Data":"53c467d6c86acab69b09453c351345abad4d242933511dafc4c690b1557b03bf"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.052307 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" event={"ID":"4a04cd83-dbd7-4076-858a-ff36748ed2d1","Type":"ContainerStarted","Data":"b20ed7e0c7c72f77a6776cc2ba8535cc357b6171111c265485609eaa79616471"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.060379 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pn9hw"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.060929 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" event={"ID":"7cf2b5e5-0623-438d-bee5-f2403d7f6e43","Type":"ContainerStarted","Data":"ae8c5d9e660ce5ca38c98bfe11e849c36847e4ca4273775486cce8641f1f388f"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.062713 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/667c43a4-aeb4-46b4-996a-e0695695dbbd-certs\") pod \"machine-config-server-8ttx7\" (UID: \"667c43a4-aeb4-46b4-996a-e0695695dbbd\") " pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.063307 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/638083de-01ec-4816-a985-db72a26ccce1-metrics-tls\") pod \"dns-default-x6xql\" (UID: \"638083de-01ec-4816-a985-db72a26ccce1\") " pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.064075 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0b2bce21-b555-4085-9d07-a1da0661bab6-profile-collector-cert\") pod \"catalog-operator-68c6474976-b9zwp\" (UID: \"0b2bce21-b555-4085-9d07-a1da0661bab6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.064423 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zqm8\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-kube-api-access-6zqm8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.067254 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xxs96\" (UID: \"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.098423 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5xgx\" (UniqueName: \"kubernetes.io/projected/d9f6eaec-4e9a-4c19-b95f-cf103db54fdc-kube-api-access-p5xgx\") pod \"kube-storage-version-migrator-operator-b67b599dd-x46h8\" (UID: \"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.105649 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" event={"ID":"021e5871-36e8-48d3-92de-e504a9429d9e","Type":"ContainerStarted","Data":"0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.105739 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" event={"ID":"021e5871-36e8-48d3-92de-e504a9429d9e","Type":"ContainerStarted","Data":"e6695130a87c43b98c1b1e876d378e44592156ecc618cb09241abcdd7b671a06"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.105793 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.107215 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.112490 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmlpv\" (UniqueName: \"kubernetes.io/projected/c48ef73d-9496-48ae-98a5-7a3fbfe60187-kube-api-access-xmlpv\") pod \"packageserver-d55dfcdfc-d6tl9\" (UID: \"c48ef73d-9496-48ae-98a5-7a3fbfe60187\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.112640 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzqnk\" (UniqueName: \"kubernetes.io/projected/0a86ffc6-40bc-4d9b-b23f-37bac308a563-kube-api-access-fzqnk\") pod \"service-ca-9c57cc56f-n9zm5\" (UID: \"0a86ffc6-40bc-4d9b-b23f-37bac308a563\") " pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.112878 4575 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-t2mds container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.112938 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" podUID="021e5871-36e8-48d3-92de-e504a9429d9e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.113001 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/667c43a4-aeb4-46b4-996a-e0695695dbbd-node-bootstrap-token\") pod \"machine-config-server-8ttx7\" (UID: \"667c43a4-aeb4-46b4-996a-e0695695dbbd\") " pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.114402 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.123850 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9r5mx"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.137307 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.139099 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.140735 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:30.640713974 +0000 UTC m=+141.969272798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.143286 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" event={"ID":"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1","Type":"ContainerStarted","Data":"6d187eb9e1e66bf40ccce61ed1e16edbf17a24402af002d105e4af9c32db4f38"} Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.154073 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt6h9\" (UniqueName: \"kubernetes.io/projected/4caea03c-cd82-4798-b8b1-7007cc09079a-kube-api-access-mt6h9\") pod \"collect-profiles-29325870-d7k6g\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.162378 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlf98\" (UniqueName: \"kubernetes.io/projected/800a54e1-61b2-4e0a-a9a4-423456d0fe78-kube-api-access-tlf98\") pod \"migrator-59844c95c7-8rt5h\" (UID: \"800a54e1-61b2-4e0a-a9a4-423456d0fe78\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.187525 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-bound-sa-token\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.199639 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.202244 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.205765 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:30 crc kubenswrapper[4575]: W1004 04:36:30.207157 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdddbf25_7363_4abe_8363_90ed28ff079e.slice/crio-fa2f2961f774f1d30bfd0581c3efefabaec45681e2f9991a595c9e0239f98302 WatchSource:0}: Error finding container fa2f2961f774f1d30bfd0581c3efefabaec45681e2f9991a595c9e0239f98302: Status 404 returned error can't find the container with id fa2f2961f774f1d30bfd0581c3efefabaec45681e2f9991a595c9e0239f98302 Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.210667 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.212340 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.217661 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.219259 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e12a025a-24e2-4860-847c-4c2ec660015c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.223905 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.228787 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldppk\" (UniqueName: \"kubernetes.io/projected/e12a025a-24e2-4860-847c-4c2ec660015c-kube-api-access-ldppk\") pod \"ingress-operator-5b745b69d9-dqnb2\" (UID: \"e12a025a-24e2-4860-847c-4c2ec660015c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.233289 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.240416 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.242747 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:30.742729536 +0000 UTC m=+142.071288460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.243024 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4mn7\" (UniqueName: \"kubernetes.io/projected/616523ee-ffd4-454a-a527-1ef2838edaef-kube-api-access-w4mn7\") pod \"ingress-canary-4fdmd\" (UID: \"616523ee-ffd4-454a-a527-1ef2838edaef\") " pod="openshift-ingress-canary/ingress-canary-4fdmd" Oct 04 04:36:30 crc kubenswrapper[4575]: W1004 04:36:30.246176 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82bbeaf3_64ae_4676_a365_d75fdf225448.slice/crio-4cbb6fa744001d835c137e1b0de8601bf2d0c80c803d5cfb2e46aa742cff2a09 WatchSource:0}: Error finding container 4cbb6fa744001d835c137e1b0de8601bf2d0c80c803d5cfb2e46aa742cff2a09: Status 404 returned error can't find the container with id 4cbb6fa744001d835c137e1b0de8601bf2d0c80c803d5cfb2e46aa742cff2a09 Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.277643 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wk2s"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.313740 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcddd\" (UniqueName: \"kubernetes.io/projected/eb94e3f3-e813-4122-8718-26d6e005c8f7-kube-api-access-tcddd\") pod \"csi-hostpathplugin-lxmpg\" (UID: \"eb94e3f3-e813-4122-8718-26d6e005c8f7\") " pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.328342 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vgclk"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.335612 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86c9t\" (UniqueName: \"kubernetes.io/projected/638083de-01ec-4816-a985-db72a26ccce1-kube-api-access-86c9t\") pod \"dns-default-x6xql\" (UID: \"638083de-01ec-4816-a985-db72a26ccce1\") " pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.341179 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.342050 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnns4\" (UniqueName: \"kubernetes.io/projected/667c43a4-aeb4-46b4-996a-e0695695dbbd-kube-api-access-mnns4\") pod \"machine-config-server-8ttx7\" (UID: \"667c43a4-aeb4-46b4-996a-e0695695dbbd\") " pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.342393 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.342674 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:30.842656737 +0000 UTC m=+142.171215561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.342768 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.343163 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:30.843153482 +0000 UTC m=+142.171712296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.349749 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14d2bb1f_8844_48cd_ba55_0782b444064e.slice/crio-conmon-d4b150c904a1b5c417aadcfd67b295ef3c025d79f1f7b3f0aac496a5d83d21fe.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.374078 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2q455"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.392266 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.444348 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.445034 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:30.945007299 +0000 UTC m=+142.273566113 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.445179 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.463733 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.517279 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.539870 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4fdmd" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.554805 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.555165 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.055150567 +0000 UTC m=+142.383709391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.576417 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.582471 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.589963 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8ttx7" Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.655470 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.656275 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.155895542 +0000 UTC m=+142.484454356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.761119 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.762378 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.262357893 +0000 UTC m=+142.590916707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.843039 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f7lqb"] Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.862444 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.863570 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.363550131 +0000 UTC m=+142.692108945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:30 crc kubenswrapper[4575]: I1004 04:36:30.965981 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:30 crc kubenswrapper[4575]: E1004 04:36:30.966328 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.466316845 +0000 UTC m=+142.794875659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.011365 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-prnp7"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.066800 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:31 crc kubenswrapper[4575]: E1004 04:36:31.071744 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.567643776 +0000 UTC m=+142.896202590 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.168522 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:31 crc kubenswrapper[4575]: E1004 04:36:31.168829 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.668813863 +0000 UTC m=+142.997372857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.184560 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" event={"ID":"cdddbf25-7363-4abe-8363-90ed28ff079e","Type":"ContainerStarted","Data":"fa2f2961f774f1d30bfd0581c3efefabaec45681e2f9991a595c9e0239f98302"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.195562 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" event={"ID":"3322e445-5258-46de-8fda-0c567244c1fa","Type":"ContainerStarted","Data":"aaff117691ee6b0e5ff6b893f427aab23b01fd8f6c7545c0808f99bfccabdaf1"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.205523 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" podStartSLOduration=122.205505602 podStartE2EDuration="2m2.205505602s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:31.204092331 +0000 UTC m=+142.532651165" watchObservedRunningTime="2025-10-04 04:36:31.205505602 +0000 UTC m=+142.534064416" Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.237466 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.278925 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" event={"ID":"7cf2b5e5-0623-438d-bee5-f2403d7f6e43","Type":"ContainerStarted","Data":"85c9ed46a55d3910250955e6d0e8501163f972541d844fd4bad1bae89564469b"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.279485 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:31 crc kubenswrapper[4575]: E1004 04:36:31.281436 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.781399383 +0000 UTC m=+143.109958387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.285004 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.287987 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" event={"ID":"e4ea004b-2e51-4b67-9238-b3fbe4575d72","Type":"ContainerStarted","Data":"8d98be36da5081a7c0f2fda8f0e83acd1c3292c2caa48a7995679ed75f83500d"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.335165 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" event={"ID":"6d607e6e-db54-4e24-aafa-8fa7c16c949b","Type":"ContainerStarted","Data":"1733cdd8e7e68bb55b0438a525dd726dd386fb33900c17d0f3b564148c7c9492"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.339682 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bd5sj" event={"ID":"32d2100b-5750-4681-a3c7-5890f03107b7","Type":"ContainerStarted","Data":"f1b2e7955f8b9a3ad1ebb2337673d6ed776384731d0133b2a470b87406368927"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.348200 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.358014 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vgclk" event={"ID":"8bf9078b-7040-4b98-86d5-1199c124ccb1","Type":"ContainerStarted","Data":"8d5595b39a358c15867d8e09b1b2d9008189812e3767a6a8e9882b0e399d82bd"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.360549 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" event={"ID":"30854660-94ff-476c-a865-d326c3874db9","Type":"ContainerStarted","Data":"7a37dc496f9688595450109b5723f0e9061211df5bba1387dedcb683cd558d58"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.378141 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.378214 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" event={"ID":"14d2bb1f-8844-48cd-ba55-0782b444064e","Type":"ContainerDied","Data":"d4b150c904a1b5c417aadcfd67b295ef3c025d79f1f7b3f0aac496a5d83d21fe"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.378213 4575 generic.go:334] "Generic (PLEG): container finished" podID="14d2bb1f-8844-48cd-ba55-0782b444064e" containerID="d4b150c904a1b5c417aadcfd67b295ef3c025d79f1f7b3f0aac496a5d83d21fe" exitCode=0 Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.380876 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:31 crc kubenswrapper[4575]: E1004 04:36:31.381257 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.881241202 +0000 UTC m=+143.209800016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.409205 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" event={"ID":"c2da083b-f94f-4b8f-9657-be52a3ad66fa","Type":"ContainerStarted","Data":"e33f7bba570fc5e8ebcdeb8ef5c6eecd9629fd5fdb9d3e918acaf6ec502d48c5"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.414453 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" event={"ID":"2fe0947a-c14b-498c-ba57-c5f8733ae76f","Type":"ContainerStarted","Data":"b56ca3feff0fa80860b00befdab814e84bc406dad442c2d0db92f9713df7c3d6"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.432219 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" event={"ID":"8b2f359d-b77c-4a0b-a1f3-ff5b8e198eb1","Type":"ContainerStarted","Data":"b934fa0a8207915c40e01440271daa93dfa6695983dd2b0dec86ed2c369565cb"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.469883 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" event={"ID":"80398e72-5e3e-44ca-a188-8fd4db2ec02e","Type":"ContainerStarted","Data":"d6cfc0babef86d2c6837777ac63dc391c7cef1a57b37f823d0248b9c67064cfd"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.480904 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" event={"ID":"952902f2-2371-4a82-84e7-365978f206f6","Type":"ContainerStarted","Data":"32e1318cf7cc1f85162a7838853aef966e4a032b09e33a1efa7a91fd69944bc9"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.488536 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:31 crc kubenswrapper[4575]: E1004 04:36:31.490177 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:31.990153084 +0000 UTC m=+143.318711898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.502822 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" event={"ID":"1510c298-3548-442a-af69-72a237934ef4","Type":"ContainerStarted","Data":"53aae59562861b08d947147217c324cf39adc4b17afaa9659c1425f5c89df8e8"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.525896 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.533693 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" event={"ID":"dbe8f63c-7cac-47fb-b961-58c717e7a951","Type":"ContainerStarted","Data":"697fbd2f912c8d8c080c378c93849c776b573465040fd4b15e3c8e4f44f4b702"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.546815 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9r5mx" event={"ID":"82bbeaf3-64ae-4676-a365-d75fdf225448","Type":"ContainerStarted","Data":"4cbb6fa744001d835c137e1b0de8601bf2d0c80c803d5cfb2e46aa742cff2a09"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.558346 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.574562 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mmjn9"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.584319 4575 generic.go:334] "Generic (PLEG): container finished" podID="88d50f89-59b7-4772-8f49-7f7bfcc787ff" containerID="eeb73c1a7d6267064804d9d5f0ad45a2bf5ca968fa28e5450c683570e6c47046" exitCode=0 Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.584422 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" event={"ID":"88d50f89-59b7-4772-8f49-7f7bfcc787ff","Type":"ContainerDied","Data":"eeb73c1a7d6267064804d9d5f0ad45a2bf5ca968fa28e5450c683570e6c47046"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.594915 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:31 crc kubenswrapper[4575]: E1004 04:36:31.595849 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:32.095831343 +0000 UTC m=+143.424390257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.606379 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" event={"ID":"289056ee-b4e0-4eda-b826-5112ba8e54c0","Type":"ContainerStarted","Data":"4e4d4299c5dc9062f7afbcefee773f7d594f1a547f2cdead9c990b1aad5b3e97"} Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.629682 4575 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-flrql container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.629764 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" podUID="c72540e8-b2dd-4377-8a17-717fcb2a05a2" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.642562 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.658842 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.682781 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.698553 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:31 crc kubenswrapper[4575]: E1004 04:36:31.703179 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:32.203160719 +0000 UTC m=+143.531719533 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.785177 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-b9fb4"] Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.802869 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:31 crc kubenswrapper[4575]: E1004 04:36:31.803233 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:32.303218574 +0000 UTC m=+143.631777388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: W1004 04:36:31.883279 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49f5d8e5_b4d2_488d_bb9f_b67e02a1f32e.slice/crio-566afd672a79d8aac5bfe5b1398d2d79e211f12b5d6d3df29f2c62840198ae26 WatchSource:0}: Error finding container 566afd672a79d8aac5bfe5b1398d2d79e211f12b5d6d3df29f2c62840198ae26: Status 404 returned error can't find the container with id 566afd672a79d8aac5bfe5b1398d2d79e211f12b5d6d3df29f2c62840198ae26 Oct 04 04:36:31 crc kubenswrapper[4575]: W1004 04:36:31.883993 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18c77677_7e1e_4194_b117_74081b151270.slice/crio-b072ef3db6b592edededf079da567e3e39e8dac092f0e85160ae3c4ea8569f99 WatchSource:0}: Error finding container b072ef3db6b592edededf079da567e3e39e8dac092f0e85160ae3c4ea8569f99: Status 404 returned error can't find the container with id b072ef3db6b592edededf079da567e3e39e8dac092f0e85160ae3c4ea8569f99 Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.908955 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:31 crc kubenswrapper[4575]: E1004 04:36:31.909400 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:32.409383967 +0000 UTC m=+143.737942771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:31 crc kubenswrapper[4575]: I1004 04:36:31.994758 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmkmw" podStartSLOduration=122.994726853 podStartE2EDuration="2m2.994726853s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:31.994490126 +0000 UTC m=+143.323048970" watchObservedRunningTime="2025-10-04 04:36:31.994726853 +0000 UTC m=+143.323285667" Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.026903 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.027341 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:32.527325112 +0000 UTC m=+143.855883926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.042946 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c54b8"] Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.079514 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-n9zm5"] Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.094891 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" podStartSLOduration=122.094754557 podStartE2EDuration="2m2.094754557s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:32.087352921 +0000 UTC m=+143.415911745" watchObservedRunningTime="2025-10-04 04:36:32.094754557 +0000 UTC m=+143.423313391" Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.134708 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.135030 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:32.635017669 +0000 UTC m=+143.963576483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.224672 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drt6k" podStartSLOduration=123.22465582 podStartE2EDuration="2m3.22465582s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:32.222723244 +0000 UTC m=+143.551282058" watchObservedRunningTime="2025-10-04 04:36:32.22465582 +0000 UTC m=+143.553214634" Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.235491 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.235866 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:32.735855636 +0000 UTC m=+144.064414450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.326054 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5v5vj" podStartSLOduration=123.326035043 podStartE2EDuration="2m3.326035043s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:32.258224468 +0000 UTC m=+143.586783302" watchObservedRunningTime="2025-10-04 04:36:32.326035043 +0000 UTC m=+143.654593867" Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.331365 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h"] Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.358481 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.359051 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:32.859035184 +0000 UTC m=+144.187593998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.408750 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fn6w5"] Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.422800 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-rdr7j" podStartSLOduration=123.422778171 podStartE2EDuration="2m3.422778171s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:32.410445052 +0000 UTC m=+143.739003876" watchObservedRunningTime="2025-10-04 04:36:32.422778171 +0000 UTC m=+143.751336995" Oct 04 04:36:32 crc kubenswrapper[4575]: W1004 04:36:32.424737 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod800a54e1_61b2_4e0a_a9a4_423456d0fe78.slice/crio-68ef15cdd9af95e79547df7eed991d958957565261e60c9ab9afdb6c4f437d7d WatchSource:0}: Error finding container 68ef15cdd9af95e79547df7eed991d958957565261e60c9ab9afdb6c4f437d7d: Status 404 returned error can't find the container with id 68ef15cdd9af95e79547df7eed991d958957565261e60c9ab9afdb6c4f437d7d Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.465655 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.466114 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:32.966097783 +0000 UTC m=+144.294656597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.572927 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.573422 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:33.073399189 +0000 UTC m=+144.401958003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.578574 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9"] Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.594226 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4fdmd"] Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.666165 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lxmpg"] Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.690560 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.691016 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:33.190943053 +0000 UTC m=+144.519501867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.702013 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" event={"ID":"205c1018-8396-442c-b265-42ee18193eba","Type":"ContainerStarted","Data":"0cdf940e2746d4763eab2cb233aa7cf6324c3565599be8e85c2d7f4451a7b246"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.719069 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8ttx7" event={"ID":"667c43a4-aeb4-46b4-996a-e0695695dbbd","Type":"ContainerStarted","Data":"e5176994bac92e373e8012fd8ab89e4f2c2c7be17ca8397e71b16eb7d46eb292"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.768015 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" event={"ID":"8bf4da93-109f-4201-8fde-205d3606733f","Type":"ContainerStarted","Data":"2ce1034283a3371a1c4c38f97aa70b4ca4818073c55ed9530cead6a99e5b34c3"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.793779 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.794200 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:33.294185301 +0000 UTC m=+144.622744115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.799720 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" event={"ID":"dbe8f63c-7cac-47fb-b961-58c717e7a951","Type":"ContainerStarted","Data":"668a92ae20a1dbf0fecae2695e97a4c1a7cc53d2ed5dbbc857cc797e61f95677"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.807014 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" event={"ID":"e12a025a-24e2-4860-847c-4c2ec660015c","Type":"ContainerStarted","Data":"827f8e217038a7b2f1c1bd6d44bd3478c4bb65df3cb2acbfff6bf189d5560872"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.822767 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" event={"ID":"e4ea004b-2e51-4b67-9238-b3fbe4575d72","Type":"ContainerStarted","Data":"bd4f07d87085b056f36465abbbe9e0c17b4536c0fe22a4b29182bfe364d2f908"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.823400 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.825262 4575 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-4x8qb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.825303 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" podUID="e4ea004b-2e51-4b67-9238-b3fbe4575d72" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.855604 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" podStartSLOduration=122.855565839 podStartE2EDuration="2m2.855565839s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:32.853021155 +0000 UTC m=+144.181579979" watchObservedRunningTime="2025-10-04 04:36:32.855565839 +0000 UTC m=+144.184124653" Oct 04 04:36:32 crc kubenswrapper[4575]: W1004 04:36:32.895665 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod616523ee_ffd4_454a_a527_1ef2838edaef.slice/crio-9bb08cee7a147abb5bfda79a1aa64a5e45d2a6e15272cbf862667df758a0e31d WatchSource:0}: Error finding container 9bb08cee7a147abb5bfda79a1aa64a5e45d2a6e15272cbf862667df758a0e31d: Status 404 returned error can't find the container with id 9bb08cee7a147abb5bfda79a1aa64a5e45d2a6e15272cbf862667df758a0e31d Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.896834 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.897126 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:33.397114009 +0000 UTC m=+144.725672823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.908023 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" event={"ID":"0a86ffc6-40bc-4d9b-b23f-37bac308a563","Type":"ContainerStarted","Data":"c87d87182724b1d63488385845435f6b8752ae4fd8ec79e5a3b802b237d02e03"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.936537 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" event={"ID":"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc","Type":"ContainerStarted","Data":"58adaf1101c554300a10299f38d3df262406fa9f98a7a05ccd190ac3ed87ec6b"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.947129 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" event={"ID":"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e","Type":"ContainerStarted","Data":"566afd672a79d8aac5bfe5b1398d2d79e211f12b5d6d3df29f2c62840198ae26"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.949396 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" event={"ID":"4caea03c-cd82-4798-b8b1-7007cc09079a","Type":"ContainerStarted","Data":"80328fdb977e2cab5494a4cb779735ad9c90fb208b7ce3c0234565c8313c974f"} Oct 04 04:36:32 crc kubenswrapper[4575]: W1004 04:36:32.957878 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb94e3f3_e813_4122_8718_26d6e005c8f7.slice/crio-6bc48c19f1d3fb06f1f5dd3521e0db22812b8bf0d5d4abcdb2fef629b65da45b WatchSource:0}: Error finding container 6bc48c19f1d3fb06f1f5dd3521e0db22812b8bf0d5d4abcdb2fef629b65da45b: Status 404 returned error can't find the container with id 6bc48c19f1d3fb06f1f5dd3521e0db22812b8bf0d5d4abcdb2fef629b65da45b Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.980395 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" event={"ID":"4c57f905-7adb-4893-bb65-2f2f0adc8b5d","Type":"ContainerStarted","Data":"e8f94cf4dcb107dcaff8273bf4d90dc5293c769be33cd75657f1232822778488"} Oct 04 04:36:32 crc kubenswrapper[4575]: I1004 04:36:32.997835 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:32 crc kubenswrapper[4575]: E1004 04:36:32.998141 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:33.498127152 +0000 UTC m=+144.826685966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.002514 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" event={"ID":"ec8d8db8-33c4-433c-8d67-67546bd93235","Type":"ContainerStarted","Data":"0288070a5ed445fee33c919c4c6499cfeb1ce41d6fe677ba6ed3847b0f334411"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.005214 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x6xql"] Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.035429 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bd5sj" event={"ID":"32d2100b-5750-4681-a3c7-5890f03107b7","Type":"ContainerStarted","Data":"2cd404248fd6b86dba6a2bd37cf5c9123e722e2d55412618d99172998ab5775a"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.072564 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-bd5sj" podStartSLOduration=124.072533649 podStartE2EDuration="2m4.072533649s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:33.067323967 +0000 UTC m=+144.395882791" watchObservedRunningTime="2025-10-04 04:36:33.072533649 +0000 UTC m=+144.401092463" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.085839 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" event={"ID":"2fe0947a-c14b-498c-ba57-c5f8733ae76f","Type":"ContainerStarted","Data":"fd1b9e6fb65fbb6462cb78b9f808c744d82c5f0746485883e410a66a57f5cfeb"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.099255 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.100707 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:33.600694979 +0000 UTC m=+144.929253793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.141629 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h" event={"ID":"800a54e1-61b2-4e0a-a9a4-423456d0fe78","Type":"ContainerStarted","Data":"68ef15cdd9af95e79547df7eed991d958957565261e60c9ab9afdb6c4f437d7d"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.158818 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" event={"ID":"18c77677-7e1e-4194-b117-74081b151270","Type":"ContainerStarted","Data":"b072ef3db6b592edededf079da567e3e39e8dac092f0e85160ae3c4ea8569f99"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.176440 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" event={"ID":"886b072f-91e0-42a4-ac24-6ce0b25956a9","Type":"ContainerStarted","Data":"b3918da3809942e94a5975bc4586053f60b21e797a73c7953274557b340d1e53"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.190793 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" event={"ID":"cdddbf25-7363-4abe-8363-90ed28ff079e","Type":"ContainerStarted","Data":"be64fb35dd214d97af435bb378ed821529df8e6f125b018c7f9760606381d5d4"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.207144 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.207694 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:33.707674326 +0000 UTC m=+145.036233140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.212716 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" event={"ID":"1510c298-3548-442a-af69-72a237934ef4","Type":"ContainerStarted","Data":"4579a4e3369021b6456f82c4fc6e7b04ad8f0cb20f32744c7a60e5169f4bc853"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.233756 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" event={"ID":"844cd2d0-93d0-45d9-b342-b86df49548d8","Type":"ContainerStarted","Data":"5e469dfa42bf7816bd45841f333dd23174ae7e94cac2591f2a9e7032018096fb"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.236764 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vgclk" event={"ID":"8bf9078b-7040-4b98-86d5-1199c124ccb1","Type":"ContainerStarted","Data":"bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.242287 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5js5f" podStartSLOduration=124.242271014 podStartE2EDuration="2m4.242271014s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:33.107087346 +0000 UTC m=+144.435646180" watchObservedRunningTime="2025-10-04 04:36:33.242271014 +0000 UTC m=+144.570829828" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.317047 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.318242 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:33.818222926 +0000 UTC m=+145.146781790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.377087 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.377153 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.402863 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9r5mx" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.403912 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-9r5mx" podStartSLOduration=124.403890372 podStartE2EDuration="2m4.403890372s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:33.400708419 +0000 UTC m=+144.729267233" watchObservedRunningTime="2025-10-04 04:36:33.403890372 +0000 UTC m=+144.732449186" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.405238 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-vgclk" podStartSLOduration=124.405224411 podStartE2EDuration="2m4.405224411s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:33.304863347 +0000 UTC m=+144.633422171" watchObservedRunningTime="2025-10-04 04:36:33.405224411 +0000 UTC m=+144.733783235" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.419029 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" event={"ID":"0b2bce21-b555-4085-9d07-a1da0661bab6","Type":"ContainerStarted","Data":"5fe65b9292a4b6cca8fc82c9fce8739a1e5fefae88ee1c17e298a46d4c3ab35f"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.419957 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.420256 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.420419 4575 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-b9zwp container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.420537 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" podUID="0b2bce21-b555-4085-9d07-a1da0661bab6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.420723 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:33.920706322 +0000 UTC m=+145.249265136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.438093 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" event={"ID":"289056ee-b4e0-4eda-b826-5112ba8e54c0","Type":"ContainerStarted","Data":"952e797e357f39aa084025739935d6eb87c71d0674c99ce080dac050a1b012fa"} Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.467930 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" podStartSLOduration=123.467909067 podStartE2EDuration="2m3.467909067s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:33.453984791 +0000 UTC m=+144.782543615" watchObservedRunningTime="2025-10-04 04:36:33.467909067 +0000 UTC m=+144.796467881" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.495550 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gtmvv" podStartSLOduration=124.495532091 podStartE2EDuration="2m4.495532091s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:33.49479185 +0000 UTC m=+144.823350674" watchObservedRunningTime="2025-10-04 04:36:33.495532091 +0000 UTC m=+144.824090905" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.525660 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.540060 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.040040738 +0000 UTC m=+145.368599552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.628112 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.628312 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.128294279 +0000 UTC m=+145.456853083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.628581 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.628936 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.128924837 +0000 UTC m=+145.457483661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.729767 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.729920 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.229889828 +0000 UTC m=+145.558448652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.730046 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.730393 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.230381173 +0000 UTC m=+145.558939987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.830879 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.831109 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.831319 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.331245801 +0000 UTC m=+145.659804635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.831623 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.831985 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.331971712 +0000 UTC m=+145.660530526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.833077 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.833117 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.932920 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.933281 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.433249782 +0000 UTC m=+145.761808596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:33 crc kubenswrapper[4575]: I1004 04:36:33.933657 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:33 crc kubenswrapper[4575]: E1004 04:36:33.934084 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.434067476 +0000 UTC m=+145.762626290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.034328 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.034999 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.534979056 +0000 UTC m=+145.863537870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.136102 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.136790 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.636774801 +0000 UTC m=+145.965333615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.238015 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.238655 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.738630058 +0000 UTC m=+146.067188872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.339866 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.340562 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.840534207 +0000 UTC m=+146.169093021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.441530 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.441765 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.941721214 +0000 UTC m=+146.270280028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.442153 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.442777 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:34.942764445 +0000 UTC m=+146.271323259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.445549 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" event={"ID":"14d2bb1f-8844-48cd-ba55-0782b444064e","Type":"ContainerStarted","Data":"7197ce2539a32587207a0f4b6fb2c07acd470e6d70267b3858655a93a6183844"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.447383 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" event={"ID":"7cf2b5e5-0623-438d-bee5-f2403d7f6e43","Type":"ContainerStarted","Data":"a450c582a14644ae01c3f903248c739239b8e81cb435130499053ba7d9e8c51c"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.448309 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4fdmd" event={"ID":"616523ee-ffd4-454a-a527-1ef2838edaef","Type":"ContainerStarted","Data":"9bb08cee7a147abb5bfda79a1aa64a5e45d2a6e15272cbf862667df758a0e31d"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.450061 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" event={"ID":"8bf4da93-109f-4201-8fde-205d3606733f","Type":"ContainerStarted","Data":"02dafe222dc08ef6b4f4764cacfeb448dae21b87b103194e5371117b05778aff"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.451926 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" event={"ID":"952902f2-2371-4a82-84e7-365978f206f6","Type":"ContainerStarted","Data":"7a921e49a9dafc55a414db61be8942c3d244f9c0754ff932610c9a895d17b23a"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.453182 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.454363 4575 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2wk2s container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.454438 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" podUID="952902f2-2371-4a82-84e7-365978f206f6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.454391 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x6xql" event={"ID":"638083de-01ec-4816-a985-db72a26ccce1","Type":"ContainerStarted","Data":"15883996ca5a3d8bbe691af65a0663a70b3563a68b6d46ca0e2c771b13315a48"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.456256 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" event={"ID":"c2da083b-f94f-4b8f-9657-be52a3ad66fa","Type":"ContainerStarted","Data":"7e39ebfd6794abbeef933303b920575df998713a9388e39642561cffe76dd372"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.457748 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9r5mx" event={"ID":"82bbeaf3-64ae-4676-a365-d75fdf225448","Type":"ContainerStarted","Data":"48a3f37d4be0b8ed7e6877209bb19dc7ef53e56a230ef6e9edcb1a04a5394b42"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.458255 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.458356 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.460382 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" event={"ID":"e12a025a-24e2-4860-847c-4c2ec660015c","Type":"ContainerStarted","Data":"6f294f94040a6a43f0dfe67a7ffd3735b7bf8e373ed5b1ecc41c323bb1dcfb0b"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.461833 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" event={"ID":"18c77677-7e1e-4194-b117-74081b151270","Type":"ContainerStarted","Data":"0d1d93199be9251096cbe4817262e41c1b67210ba6251a15374cdf62aa4cf2f8"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.463169 4575 generic.go:334] "Generic (PLEG): container finished" podID="1510c298-3548-442a-af69-72a237934ef4" containerID="4579a4e3369021b6456f82c4fc6e7b04ad8f0cb20f32744c7a60e5169f4bc853" exitCode=0 Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.463260 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" event={"ID":"1510c298-3548-442a-af69-72a237934ef4","Type":"ContainerDied","Data":"4579a4e3369021b6456f82c4fc6e7b04ad8f0cb20f32744c7a60e5169f4bc853"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.464274 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8ttx7" event={"ID":"667c43a4-aeb4-46b4-996a-e0695695dbbd","Type":"ContainerStarted","Data":"f4bbf5ead5df4bd74f53da326046ce0c3f26e8f98b4f40f8219bab23d5230c27"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.466216 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" event={"ID":"4c57f905-7adb-4893-bb65-2f2f0adc8b5d","Type":"ContainerStarted","Data":"7bb6f57ea916d5e0fc971b0e13ced858a0f90e0f914ec3dd6765899ed6fee139"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.467564 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" event={"ID":"0b2bce21-b555-4085-9d07-a1da0661bab6","Type":"ContainerStarted","Data":"7a37ec07d3af50608f0170c4fc4b9c46f3b7cf5c174c891021cebd51730b668b"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.471229 4575 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-b9zwp container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.471356 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" podUID="0b2bce21-b555-4085-9d07-a1da0661bab6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.475207 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" event={"ID":"ec8d8db8-33c4-433c-8d67-67546bd93235","Type":"ContainerStarted","Data":"8abc7679f2e0b4f8fd8e2781e6cb69a0c8b17e570fd5748a3b67b1295306ef53"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.493220 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" event={"ID":"eb94e3f3-e813-4122-8718-26d6e005c8f7","Type":"ContainerStarted","Data":"6bc48c19f1d3fb06f1f5dd3521e0db22812b8bf0d5d4abcdb2fef629b65da45b"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.503974 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" event={"ID":"30854660-94ff-476c-a865-d326c3874db9","Type":"ContainerStarted","Data":"7e7b22a7b8798059429199d01a6326e7df7d5d0f5d7fb20fe8d0cb3f922b5c5d"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.504542 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.505626 4575 patch_prober.go:28] interesting pod/console-operator-58897d9998-f7lqb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.505732 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" podUID="30854660-94ff-476c-a865-d326c3874db9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.516875 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" event={"ID":"cdddbf25-7363-4abe-8363-90ed28ff079e","Type":"ContainerStarted","Data":"4ce97c483381417eb050e6ddc66333cfc52e4ec21a3061a9771dfa7bb87780af"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.523185 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" event={"ID":"c48ef73d-9496-48ae-98a5-7a3fbfe60187","Type":"ContainerStarted","Data":"c26a5724aea07b4e48ed89e44afbb90c22605e03a2ab910eedb19052c655c225"} Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.525101 4575 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-4x8qb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.525281 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" podUID="e4ea004b-2e51-4b67-9238-b3fbe4575d72" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.546123 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.546300 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.04627209 +0000 UTC m=+146.374830904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.547966 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.548466 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.048453883 +0000 UTC m=+146.377012687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.574268 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" podStartSLOduration=124.574237474 podStartE2EDuration="2m4.574237474s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:34.570344301 +0000 UTC m=+145.898903135" watchObservedRunningTime="2025-10-04 04:36:34.574237474 +0000 UTC m=+145.902796288" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.650169 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.650785 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.150750373 +0000 UTC m=+146.479309227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.658613 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qmhlx" podStartSLOduration=124.658578281 podStartE2EDuration="2m4.658578281s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:34.622648175 +0000 UTC m=+145.951206999" watchObservedRunningTime="2025-10-04 04:36:34.658578281 +0000 UTC m=+145.987137095" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.694473 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h4kkh" podStartSLOduration=125.694451316 podStartE2EDuration="2m5.694451316s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:34.660909159 +0000 UTC m=+145.989467993" watchObservedRunningTime="2025-10-04 04:36:34.694451316 +0000 UTC m=+146.023010150" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.745323 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-mmjn9" podStartSLOduration=125.745304568 podStartE2EDuration="2m5.745304568s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:34.697797594 +0000 UTC m=+146.026356418" watchObservedRunningTime="2025-10-04 04:36:34.745304568 +0000 UTC m=+146.073863402" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.756959 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.757323 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.257307498 +0000 UTC m=+146.585866312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.782718 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" podStartSLOduration=125.782699327 podStartE2EDuration="2m5.782699327s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:34.746298597 +0000 UTC m=+146.074857421" watchObservedRunningTime="2025-10-04 04:36:34.782699327 +0000 UTC m=+146.111258151" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.805010 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" podStartSLOduration=125.804992397 podStartE2EDuration="2m5.804992397s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:34.784962633 +0000 UTC m=+146.113521447" watchObservedRunningTime="2025-10-04 04:36:34.804992397 +0000 UTC m=+146.133551231" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.848612 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:34 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:34 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:34 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.849075 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.858296 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.858718 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.358695701 +0000 UTC m=+146.687254525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:34 crc kubenswrapper[4575]: I1004 04:36:34.961527 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:34 crc kubenswrapper[4575]: E1004 04:36:34.962038 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.462021141 +0000 UTC m=+146.790579945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.062667 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.062959 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.56292677 +0000 UTC m=+146.891485594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.063907 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.064342 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.564326101 +0000 UTC m=+146.892884925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.164988 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.165203 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.665187019 +0000 UTC m=+146.993745833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.166728 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.167224 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.667213118 +0000 UTC m=+146.995771932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.267806 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.268183 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.768165809 +0000 UTC m=+147.096724623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.369200 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.369812 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.86979625 +0000 UTC m=+147.198355064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.470503 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.470619 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.970602636 +0000 UTC m=+147.299161440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.470758 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.471063 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:35.971052509 +0000 UTC m=+147.299611333 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.571442 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.572154 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.072134134 +0000 UTC m=+147.400692948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.579162 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" event={"ID":"886b072f-91e0-42a4-ac24-6ce0b25956a9","Type":"ContainerStarted","Data":"cb0c89c6cf4a37bb4ed039d491b30600ffeee81d589631af6c67bca8561e0c3f"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.586403 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" event={"ID":"49f5d8e5-b4d2-488d-bb9f-b67e02a1f32e","Type":"ContainerStarted","Data":"e10d29b2262e4e6cdf96300a21478e9b96f367548e9f7924df12fa8f09298cca"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.617872 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" event={"ID":"4caea03c-cd82-4798-b8b1-7007cc09079a","Type":"ContainerStarted","Data":"304fb6501316231bf2fa86a48b13d1c20f8edf18912b5bae16f7824e5a3067df"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.624900 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8ttx7" podStartSLOduration=8.6248825 podStartE2EDuration="8.6248825s" podCreationTimestamp="2025-10-04 04:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:34.807425287 +0000 UTC m=+146.135984101" watchObservedRunningTime="2025-10-04 04:36:35.6248825 +0000 UTC m=+146.953441314" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.626140 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c54b8" podStartSLOduration=125.626133887 podStartE2EDuration="2m5.626133887s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:35.625010854 +0000 UTC m=+146.953569658" watchObservedRunningTime="2025-10-04 04:36:35.626133887 +0000 UTC m=+146.954692721" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.636243 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" event={"ID":"ec8d8db8-33c4-433c-8d67-67546bd93235","Type":"ContainerStarted","Data":"6a905765d3d416d4335479d1269a9031fd166be8c8e0acd7fcc18d2616340a5b"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.636955 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.675703 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.678411 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.178399679 +0000 UTC m=+147.506958493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.688084 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" event={"ID":"88d50f89-59b7-4772-8f49-7f7bfcc787ff","Type":"ContainerStarted","Data":"a80db52d0e61587e73fb9991942ac43f5bf8623516519ed61db71dcb7227020d"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.714459 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" event={"ID":"1510c298-3548-442a-af69-72a237934ef4","Type":"ContainerStarted","Data":"c10a1d673094886529c2ead890b93191e41c0446c605586020fceca5c9fbdfdf"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.715060 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.737922 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xxs96" podStartSLOduration=126.737904032 podStartE2EDuration="2m6.737904032s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:35.70415716 +0000 UTC m=+147.032715994" watchObservedRunningTime="2025-10-04 04:36:35.737904032 +0000 UTC m=+147.066462846" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.739229 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" podStartSLOduration=126.73922156 podStartE2EDuration="2m6.73922156s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:35.737304994 +0000 UTC m=+147.065863818" watchObservedRunningTime="2025-10-04 04:36:35.73922156 +0000 UTC m=+147.067780374" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.739373 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" event={"ID":"205c1018-8396-442c-b265-42ee18193eba","Type":"ContainerStarted","Data":"e1b6b8accd1168b03178dc8d9fe5dbd3ac40c85c626730dd46a28796cd3c039b"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.765415 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" event={"ID":"dbe8f63c-7cac-47fb-b961-58c717e7a951","Type":"ContainerStarted","Data":"f9b11a222c8400aa6ba2993c5b3a8248165eb0411f62cbf4b7e9cf2cff41af3a"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.776230 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.776411 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.276393633 +0000 UTC m=+147.604952447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.776576 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.777146 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.277131025 +0000 UTC m=+147.605689839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.784443 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h" event={"ID":"800a54e1-61b2-4e0a-a9a4-423456d0fe78","Type":"ContainerStarted","Data":"c97eb5c40b301e1205a99d634c8e9e7e63003dac84ac5997c61bafa9cc9b93bd"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.797889 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" event={"ID":"18c77677-7e1e-4194-b117-74081b151270","Type":"ContainerStarted","Data":"d20c5bffe01d10d00f994fdbcfc7aa6c8070c43e8f9a0432e36d827854595377"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.840071 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" podStartSLOduration=126.840056418 podStartE2EDuration="2m6.840056418s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:35.801397421 +0000 UTC m=+147.129956235" watchObservedRunningTime="2025-10-04 04:36:35.840056418 +0000 UTC m=+147.168615232" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.857028 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" event={"ID":"d9f6eaec-4e9a-4c19-b95f-cf103db54fdc","Type":"ContainerStarted","Data":"194bfa101383c188faa60f3605d283dcebec1ef93fbddaff8ac2f5265c5215ef"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.860848 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:35 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:35 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:35 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.860901 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.881175 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" event={"ID":"80398e72-5e3e-44ca-a188-8fd4db2ec02e","Type":"ContainerStarted","Data":"9c6a997383621a10eaa0906fb1cc3b6ed38edddd0cbe1e26e0399cea19b3177f"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.881312 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:35 crc kubenswrapper[4575]: E1004 04:36:35.882568 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.382553436 +0000 UTC m=+147.711112250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.890164 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" podStartSLOduration=125.890141077 podStartE2EDuration="2m5.890141077s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:35.841364256 +0000 UTC m=+147.169923070" watchObservedRunningTime="2025-10-04 04:36:35.890141077 +0000 UTC m=+147.218699891" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.890600 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-pn9hw" podStartSLOduration=126.89059555 podStartE2EDuration="2m6.89059555s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:35.884164013 +0000 UTC m=+147.212722827" watchObservedRunningTime="2025-10-04 04:36:35.89059555 +0000 UTC m=+147.219154364" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.941940 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dsxgt" podStartSLOduration=125.941918885 podStartE2EDuration="2m5.941918885s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:35.930959006 +0000 UTC m=+147.259517820" watchObservedRunningTime="2025-10-04 04:36:35.941918885 +0000 UTC m=+147.270477699" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.965425 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4fdmd" event={"ID":"616523ee-ffd4-454a-a527-1ef2838edaef","Type":"ContainerStarted","Data":"09baa12899f77295d4f9ad1897530fef8f2b90de63e928fda5355b08ce26cefd"} Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.982712 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:35 crc kubenswrapper[4575]: I1004 04:36:35.983422 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x46h8" podStartSLOduration=126.983405284 podStartE2EDuration="2m6.983405284s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:35.975351429 +0000 UTC m=+147.303910243" watchObservedRunningTime="2025-10-04 04:36:35.983405284 +0000 UTC m=+147.311964098" Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:35.983946 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.483923989 +0000 UTC m=+147.812482893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:35.983971 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" event={"ID":"c48ef73d-9496-48ae-98a5-7a3fbfe60187","Type":"ContainerStarted","Data":"248ab89857a60248201e6704e89cf5383518f01ce826862e2df84f37e6300a61"} Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.010470 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.010578 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.010670 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" event={"ID":"844cd2d0-93d0-45d9-b342-b86df49548d8","Type":"ContainerStarted","Data":"280ba2abaf1b19ae3d72e9df4ce5ac0e9096e9de2239b8ee86023490b914e23b"} Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.010230 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.012240 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.010267 4575 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-b9zwp container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.012570 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" podUID="0b2bce21-b555-4085-9d07-a1da0661bab6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.001515 4575 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fn6w5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:35.986098 4575 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d6tl9 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.012704 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" podUID="c48ef73d-9496-48ae-98a5-7a3fbfe60187" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.012643 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" podUID="844cd2d0-93d0-45d9-b342-b86df49548d8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.011308 4575 patch_prober.go:28] interesting pod/console-operator-58897d9998-f7lqb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.012827 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" podUID="30854660-94ff-476c-a865-d326c3874db9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.013668 4575 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2wk2s container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.013701 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" podUID="952902f2-2371-4a82-84e7-365978f206f6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.010742 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x6xql" event={"ID":"638083de-01ec-4816-a985-db72a26ccce1","Type":"ContainerStarted","Data":"2aa79ebeabf4f9244cc850834c697102869f8e6ff1bdb7d24259b006cc14a4b3"} Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.015132 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" event={"ID":"e12a025a-24e2-4860-847c-4c2ec660015c","Type":"ContainerStarted","Data":"41c57363cacca2159707712ed93ffe6ec226ccf3adaa6a9086f3872914b79239"} Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.015203 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" event={"ID":"0a86ffc6-40bc-4d9b-b23f-37bac308a563","Type":"ContainerStarted","Data":"08201b01eae8987f71b124e7c96037fa4134e67ea22919865cb8afbf4bcccef6"} Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.037578 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-4fdmd" podStartSLOduration=9.037557991 podStartE2EDuration="9.037557991s" podCreationTimestamp="2025-10-04 04:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:36.037113338 +0000 UTC m=+147.365672172" watchObservedRunningTime="2025-10-04 04:36:36.037557991 +0000 UTC m=+147.366116805" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.069556 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4x8qb" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.085171 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.087488 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.587467825 +0000 UTC m=+147.916026649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.098913 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" podStartSLOduration=126.098887238 podStartE2EDuration="2m6.098887238s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:36.09346556 +0000 UTC m=+147.422024374" watchObservedRunningTime="2025-10-04 04:36:36.098887238 +0000 UTC m=+147.427446052" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.152027 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-n9zm5" podStartSLOduration=126.152012165 podStartE2EDuration="2m6.152012165s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:36.151238973 +0000 UTC m=+147.479797797" watchObservedRunningTime="2025-10-04 04:36:36.152012165 +0000 UTC m=+147.480570979" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.187710 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.188201 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.688184319 +0000 UTC m=+148.016743133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.216320 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" podStartSLOduration=126.216300808 podStartE2EDuration="2m6.216300808s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:36.183116021 +0000 UTC m=+147.511674835" watchObservedRunningTime="2025-10-04 04:36:36.216300808 +0000 UTC m=+147.544859622" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.256386 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dqnb2" podStartSLOduration=127.256350695 podStartE2EDuration="2m7.256350695s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:36.253922644 +0000 UTC m=+147.582481458" watchObservedRunningTime="2025-10-04 04:36:36.256350695 +0000 UTC m=+147.584909509" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.259188 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-b9fb4" podStartSLOduration=127.259179037 podStartE2EDuration="2m7.259179037s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:36.218014018 +0000 UTC m=+147.546572852" watchObservedRunningTime="2025-10-04 04:36:36.259179037 +0000 UTC m=+147.587737851" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.289538 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.290081 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.790052516 +0000 UTC m=+148.118611330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.320532 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q6xg8" podStartSLOduration=127.320500423 podStartE2EDuration="2m7.320500423s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:36.288828191 +0000 UTC m=+147.617387005" watchObservedRunningTime="2025-10-04 04:36:36.320500423 +0000 UTC m=+147.649059257" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.391308 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.391738 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.891717448 +0000 UTC m=+148.220276262 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.492214 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.492386 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.99235886 +0000 UTC m=+148.320917674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.492527 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.492936 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:36.992923006 +0000 UTC m=+148.321481830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.593866 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.594220 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.094203366 +0000 UTC m=+148.422762180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.594445 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.594848 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.094837395 +0000 UTC m=+148.423396209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.695705 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.695917 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.195892399 +0000 UTC m=+148.524451213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.696178 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.696532 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.196515677 +0000 UTC m=+148.525074561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.797428 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.797630 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.297600381 +0000 UTC m=+148.626159215 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.797911 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.798293 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.298281911 +0000 UTC m=+148.626840805 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.824101 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.825081 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.838969 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.841043 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.850165 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:36 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:36 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:36 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.850695 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.854809 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.899527 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.899814 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:36 crc kubenswrapper[4575]: I1004 04:36:36.899899 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:36 crc kubenswrapper[4575]: E1004 04:36:36.900128 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.400107158 +0000 UTC m=+148.728665972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.001451 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.001782 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.001607 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.002009 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.002055 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.502033267 +0000 UTC m=+148.830592081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.013607 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" event={"ID":"80398e72-5e3e-44ca-a188-8fd4db2ec02e","Type":"ContainerStarted","Data":"bd6908cb009be87f5dae1f4e89ebd80509fee682db4bb4c10d28b3d1ecc2a268"} Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.016619 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x6xql" event={"ID":"638083de-01ec-4816-a985-db72a26ccce1","Type":"ContainerStarted","Data":"d6e113a2086395a6945df5d2d24e7ad02235ee0e04efc6e7ceb6505a83ffe14b"} Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.017045 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.019833 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" event={"ID":"88d50f89-59b7-4772-8f49-7f7bfcc787ff","Type":"ContainerStarted","Data":"5da4b628ff45f2ffe2dfb81db8d3c1514f651b3bbcf35149b3c713f00e537777"} Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.023314 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h" event={"ID":"800a54e1-61b2-4e0a-a9a4-423456d0fe78","Type":"ContainerStarted","Data":"8710c3cd131352c035a619f512bc29b0084c3af121186f74914255c3ee9c9258"} Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.025292 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" event={"ID":"205c1018-8396-442c-b265-42ee18193eba","Type":"ContainerStarted","Data":"af9049d5d2ebc4a51ab18fe32ac186c8aa4b46a59dd79a6d046dc31ac1b5ca69"} Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.025967 4575 patch_prober.go:28] interesting pod/console-operator-58897d9998-f7lqb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.025985 4575 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fn6w5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.026007 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" podUID="30854660-94ff-476c-a865-d326c3874db9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.026010 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" podUID="844cd2d0-93d0-45d9-b342-b86df49548d8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.027080 4575 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d6tl9 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.027202 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" podUID="c48ef73d-9496-48ae-98a5-7a3fbfe60187" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.040708 4575 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-2q455 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.040780 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" podUID="1510c298-3548-442a-af69-72a237934ef4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.046288 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.072680 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-prnp7" podStartSLOduration=127.072654994 podStartE2EDuration="2m7.072654994s" podCreationTimestamp="2025-10-04 04:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:37.064356202 +0000 UTC m=+148.392915026" watchObservedRunningTime="2025-10-04 04:36:37.072654994 +0000 UTC m=+148.401213828" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.104111 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.105882 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.605866801 +0000 UTC m=+148.934425615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.107713 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.111473 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.611457874 +0000 UTC m=+148.940016688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.142170 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" podStartSLOduration=128.142148968 podStartE2EDuration="2m8.142148968s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:37.107504679 +0000 UTC m=+148.436063513" watchObservedRunningTime="2025-10-04 04:36:37.142148968 +0000 UTC m=+148.470707782" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.143294 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.143457 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-x6xql" podStartSLOduration=10.143450116 podStartE2EDuration="10.143450116s" podCreationTimestamp="2025-10-04 04:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:37.135191976 +0000 UTC m=+148.463750800" watchObservedRunningTime="2025-10-04 04:36:37.143450116 +0000 UTC m=+148.472008940" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.192497 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cdq5t" podStartSLOduration=128.192475574 podStartE2EDuration="2m8.192475574s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:37.191808295 +0000 UTC m=+148.520367109" watchObservedRunningTime="2025-10-04 04:36:37.192475574 +0000 UTC m=+148.521034398" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.209127 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.209384 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.209422 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.209541 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.709526721 +0000 UTC m=+149.038085535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.222619 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.253111 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.312202 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.312270 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.312356 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.312685 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.812668206 +0000 UTC m=+149.141227030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.328723 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.336340 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.338234 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.343871 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.414113 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.414490 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:37.914473371 +0000 UTC m=+149.243032185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.422035 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.516242 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.516563 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.016552065 +0000 UTC m=+149.345110879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.618164 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.618559 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.118539676 +0000 UTC m=+149.447098490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.719486 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.719777 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.219764585 +0000 UTC m=+149.548323399 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.733821 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8rt5h" podStartSLOduration=128.733801864 podStartE2EDuration="2m8.733801864s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:37.222851279 +0000 UTC m=+148.551410103" watchObservedRunningTime="2025-10-04 04:36:37.733801864 +0000 UTC m=+149.062360678" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.734189 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.821248 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.821470 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.321450077 +0000 UTC m=+149.650008891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.821832 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.822094 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.322086095 +0000 UTC m=+149.650644909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.835771 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:37 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:37 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:37 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.835820 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:37 crc kubenswrapper[4575]: I1004 04:36:37.922689 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:37 crc kubenswrapper[4575]: E1004 04:36:37.923045 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.423026146 +0000 UTC m=+149.751584960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.028182 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.028196 4575 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2wk2s container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.028575 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.52856401 +0000 UTC m=+149.857122824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.028567 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" podUID="952902f2-2371-4a82-84e7-365978f206f6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.104422 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" event={"ID":"eb94e3f3-e813-4122-8718-26d6e005c8f7","Type":"ContainerStarted","Data":"e2fc6337be460150a5af64eebe660f3af1f3d4c88534b47bc5a2a16db9385741"} Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.105845 4575 generic.go:334] "Generic (PLEG): container finished" podID="4caea03c-cd82-4798-b8b1-7007cc09079a" containerID="304fb6501316231bf2fa86a48b13d1c20f8edf18912b5bae16f7824e5a3067df" exitCode=0 Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.105894 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" event={"ID":"4caea03c-cd82-4798-b8b1-7007cc09079a","Type":"ContainerDied","Data":"304fb6501316231bf2fa86a48b13d1c20f8edf18912b5bae16f7824e5a3067df"} Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.107774 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4b664ad9-7e4d-4330-aa05-620e9617d1c1","Type":"ContainerStarted","Data":"c3de70e669c6d1d94db6ee26d0ea6f323e5b0a3d06ae246de82aa5e4d2147188"} Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.109140 4575 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-2q455 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.109195 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" podUID="1510c298-3548-442a-af69-72a237934ef4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.129152 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.129600 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.629563362 +0000 UTC m=+149.958122176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.230545 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.239949 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.739931467 +0000 UTC m=+150.068490281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.331286 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.331658 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.831643479 +0000 UTC m=+150.160202293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.432208 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.432476 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:38.932465156 +0000 UTC m=+150.261023970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.453371 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.453430 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.533170 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.533937 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.033906561 +0000 UTC m=+150.362465375 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.634258 4575 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-2q455 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.634313 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" podUID="1510c298-3548-442a-af69-72a237934ef4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.634568 4575 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-2q455 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.634647 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" podUID="1510c298-3548-442a-af69-72a237934ef4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.635126 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.635389 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.135379017 +0000 UTC m=+150.463937821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.743832 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.744221 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.244206427 +0000 UTC m=+150.572765241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.758636 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mkz66"] Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.774709 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.778761 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.826203 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mkz66"] Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.843055 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:38 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:38 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:38 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.843104 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.848373 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-catalog-content\") pod \"certified-operators-mkz66\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.848437 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfz6h\" (UniqueName: \"kubernetes.io/projected/7846a3af-bfa1-4129-ae19-48bc580c0a4a-kube-api-access-nfz6h\") pod \"certified-operators-mkz66\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.848485 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-utilities\") pod \"certified-operators-mkz66\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.848522 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.848835 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.348822625 +0000 UTC m=+150.677381439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.949065 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.949221 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfz6h\" (UniqueName: \"kubernetes.io/projected/7846a3af-bfa1-4129-ae19-48bc580c0a4a-kube-api-access-nfz6h\") pod \"certified-operators-mkz66\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.949270 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-utilities\") pod \"certified-operators-mkz66\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.949356 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-catalog-content\") pod \"certified-operators-mkz66\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.949791 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-catalog-content\") pod \"certified-operators-mkz66\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:38 crc kubenswrapper[4575]: E1004 04:36:38.949862 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.449848388 +0000 UTC m=+150.778407202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.950380 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-utilities\") pod \"certified-operators-mkz66\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:38 crc kubenswrapper[4575]: I1004 04:36:38.951358 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.003788 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfz6h\" (UniqueName: \"kubernetes.io/projected/7846a3af-bfa1-4129-ae19-48bc580c0a4a-kube-api-access-nfz6h\") pod \"certified-operators-mkz66\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.050728 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.051114 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.551097997 +0000 UTC m=+150.879656811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.101202 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.148168 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.148204 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.155075 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.155558 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.65554 +0000 UTC m=+150.984098814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.156324 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ec7609d921708f6c505577537ac6076b27a661bda5864707860666bdafef236e"} Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.156371 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fdba17961e830cda3f143b491caf314acdba440ef413f96a3f6718cb4b8f7886"} Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.168437 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dgzgr"] Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.169386 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.178045 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.178079 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.179180 4575 patch_prober.go:28] interesting pod/apiserver-76f77b778f-q9ssb container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.179221 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" podUID="88d50f89-59b7-4772-8f49-7f7bfcc787ff" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.189785 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8e80c79aa800bb0f1a07b10e6481eea7fbe023c54fcf14ce4e8a190deb39e008"} Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.189832 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"471f739d6ae2b8bf63a39fca2bb8ea1b9b5d1a59b777fcd04d669462e49bac78"} Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.193659 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.214789 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4b664ad9-7e4d-4330-aa05-620e9617d1c1","Type":"ContainerStarted","Data":"707fb87b72c56fc6f49c3f3f1b8604a6862286c35f647b6e83e2b39ab416c1d1"} Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.231977 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a496cad833ef909dc8a6d3507f25d61e4eb75a806da1cd697886513a02818bdb"} Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.232072 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2a3dba58c04c80059216a30d40dc778bce6d918e956dee63c79be19ce6acf23e"} Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.232678 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.257740 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-catalog-content\") pod \"certified-operators-dgzgr\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.257802 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx4jd\" (UniqueName: \"kubernetes.io/projected/bfad083f-22e8-4845-a327-7457ae07bfb8-kube-api-access-vx4jd\") pod \"certified-operators-dgzgr\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.257829 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-utilities\") pod \"certified-operators-dgzgr\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.257853 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.259086 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.759075476 +0000 UTC m=+151.087634290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.263063 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dgzgr"] Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.360480 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.360740 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.860714185 +0000 UTC m=+151.189272999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.360953 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.361080 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-catalog-content\") pod \"certified-operators-dgzgr\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.361142 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx4jd\" (UniqueName: \"kubernetes.io/projected/bfad083f-22e8-4845-a327-7457ae07bfb8-kube-api-access-vx4jd\") pod \"certified-operators-dgzgr\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.361190 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-utilities\") pod \"certified-operators-dgzgr\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.362483 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.862473667 +0000 UTC m=+151.191032481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.363069 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-catalog-content\") pod \"certified-operators-dgzgr\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.363432 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-utilities\") pod \"certified-operators-dgzgr\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.365829 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.365815964 podStartE2EDuration="3.365815964s" podCreationTimestamp="2025-10-04 04:36:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:39.34370354 +0000 UTC m=+150.672262354" watchObservedRunningTime="2025-10-04 04:36:39.365815964 +0000 UTC m=+150.694374778" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.408204 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx4jd\" (UniqueName: \"kubernetes.io/projected/bfad083f-22e8-4845-a327-7457ae07bfb8-kube-api-access-vx4jd\") pod \"certified-operators-dgzgr\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.434801 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.434864 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.435242 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.435263 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.462060 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.462317 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.962298815 +0000 UTC m=+151.290857629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.462426 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.462825 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:39.96280996 +0000 UTC m=+151.291368774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.503865 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.534627 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x4zr9"] Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.552566 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.557996 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.562967 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.563402 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.063385589 +0000 UTC m=+151.391944403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.565290 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x4zr9"] Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.647768 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.648126 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.656969 4575 patch_prober.go:28] interesting pod/console-f9d7485db-vgclk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.29:8443/health\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.657048 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vgclk" podUID="8bf9078b-7040-4b98-86d5-1199c124ccb1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.29:8443/health\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.657394 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.665517 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-catalog-content\") pod \"community-operators-x4zr9\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.665578 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.665643 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84dd2\" (UniqueName: \"kubernetes.io/projected/f9b027e7-fd36-43e5-80b2-53e105822d71-kube-api-access-84dd2\") pod \"community-operators-x4zr9\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.665724 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-utilities\") pod \"community-operators-x4zr9\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.666033 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.166020129 +0000 UTC m=+151.494578943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.770502 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.770875 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-utilities\") pod \"community-operators-x4zr9\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.770951 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-catalog-content\") pod \"community-operators-x4zr9\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.771116 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84dd2\" (UniqueName: \"kubernetes.io/projected/f9b027e7-fd36-43e5-80b2-53e105822d71-kube-api-access-84dd2\") pod \"community-operators-x4zr9\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.788918 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.288893899 +0000 UTC m=+151.617452763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.791565 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-catalog-content\") pod \"community-operators-x4zr9\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.799184 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-utilities\") pod \"community-operators-x4zr9\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.831630 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.840149 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k5482"] Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.841986 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84dd2\" (UniqueName: \"kubernetes.io/projected/f9b027e7-fd36-43e5-80b2-53e105822d71-kube-api-access-84dd2\") pod \"community-operators-x4zr9\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.861241 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k5482"] Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.861398 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.870110 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-f7lqb" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.872681 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:39 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:39 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:39 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.872726 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.873670 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.882119 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.382095614 +0000 UTC m=+151.710654428 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.892562 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.978825 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.978986 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.478956475 +0000 UTC m=+151.807515299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.979135 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-catalog-content\") pod \"community-operators-k5482\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.979211 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl6rp\" (UniqueName: \"kubernetes.io/projected/c5b5f157-4684-46a8-bf72-604f9429decc-kube-api-access-jl6rp\") pod \"community-operators-k5482\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.979344 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-utilities\") pod \"community-operators-k5482\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:39 crc kubenswrapper[4575]: I1004 04:36:39.979376 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:39 crc kubenswrapper[4575]: E1004 04:36:39.979754 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.479736568 +0000 UTC m=+151.808295382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.080169 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.080483 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-utilities\") pod \"community-operators-k5482\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.080581 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-catalog-content\") pod \"community-operators-k5482\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.080650 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl6rp\" (UniqueName: \"kubernetes.io/projected/c5b5f157-4684-46a8-bf72-604f9429decc-kube-api-access-jl6rp\") pod \"community-operators-k5482\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.085151 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.585129218 +0000 UTC m=+151.913688032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.085279 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.085812 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-utilities\") pod \"community-operators-k5482\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.086091 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.088556 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-catalog-content\") pod \"community-operators-k5482\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.105352 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.105535 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.110045 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.131637 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl6rp\" (UniqueName: \"kubernetes.io/projected/c5b5f157-4684-46a8-bf72-604f9429decc-kube-api-access-jl6rp\") pod \"community-operators-k5482\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.144347 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-b9zwp" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.183343 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.183727 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.183749 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.185505 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.685493492 +0000 UTC m=+152.014052306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.225774 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k5482" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.226820 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.227348 4575 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fn6w5 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.227383 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" podUID="844cd2d0-93d0-45d9-b342-b86df49548d8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.227566 4575 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fn6w5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.227721 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" podUID="844cd2d0-93d0-45d9-b342-b86df49548d8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.248781 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d6tl9" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.273439 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.274329 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g" event={"ID":"4caea03c-cd82-4798-b8b1-7007cc09079a","Type":"ContainerDied","Data":"80328fdb977e2cab5494a4cb779735ad9c90fb208b7ce3c0234565c8313c974f"} Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.274441 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80328fdb977e2cab5494a4cb779735ad9c90fb208b7ce3c0234565c8313c974f" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.285902 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4caea03c-cd82-4798-b8b1-7007cc09079a-secret-volume\") pod \"4caea03c-cd82-4798-b8b1-7007cc09079a\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.285964 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4caea03c-cd82-4798-b8b1-7007cc09079a-config-volume\") pod \"4caea03c-cd82-4798-b8b1-7007cc09079a\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.285994 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt6h9\" (UniqueName: \"kubernetes.io/projected/4caea03c-cd82-4798-b8b1-7007cc09079a-kube-api-access-mt6h9\") pod \"4caea03c-cd82-4798-b8b1-7007cc09079a\" (UID: \"4caea03c-cd82-4798-b8b1-7007cc09079a\") " Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.286152 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.286421 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.286444 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.288960 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4caea03c-cd82-4798-b8b1-7007cc09079a-config-volume" (OuterVolumeSpecName: "config-volume") pod "4caea03c-cd82-4798-b8b1-7007cc09079a" (UID: "4caea03c-cd82-4798-b8b1-7007cc09079a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.290787 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.790760158 +0000 UTC m=+152.119318972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.291155 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.294991 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jlb6g" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.298873 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4caea03c-cd82-4798-b8b1-7007cc09079a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4caea03c-cd82-4798-b8b1-7007cc09079a" (UID: "4caea03c-cd82-4798-b8b1-7007cc09079a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.305705 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4caea03c-cd82-4798-b8b1-7007cc09079a-kube-api-access-mt6h9" (OuterVolumeSpecName: "kube-api-access-mt6h9") pod "4caea03c-cd82-4798-b8b1-7007cc09079a" (UID: "4caea03c-cd82-4798-b8b1-7007cc09079a"). InnerVolumeSpecName "kube-api-access-mt6h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.346564 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.387814 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.388414 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt6h9\" (UniqueName: \"kubernetes.io/projected/4caea03c-cd82-4798-b8b1-7007cc09079a-kube-api-access-mt6h9\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.388541 4575 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4caea03c-cd82-4798-b8b1-7007cc09079a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.388652 4575 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4caea03c-cd82-4798-b8b1-7007cc09079a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.392005 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.891987067 +0000 UTC m=+152.220545881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.478942 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.483873 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mkz66"] Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.490762 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.491018 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.990997871 +0000 UTC m=+152.319556695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.491507 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.491968 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:40.991958369 +0000 UTC m=+152.320517183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.593837 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.594231 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:41.094216208 +0000 UTC m=+152.422775022 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.695683 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.696073 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:41.196061155 +0000 UTC m=+152.524619969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.797118 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.797656 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:41.297640524 +0000 UTC m=+152.626199328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.844818 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:40 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:40 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:40 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.844873 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.901361 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:40 crc kubenswrapper[4575]: E1004 04:36:40.901704 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:41.401672375 +0000 UTC m=+152.730231189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:40 crc kubenswrapper[4575]: I1004 04:36:40.991494 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dgzgr"] Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.002317 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.002707 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:41.502692168 +0000 UTC m=+152.831250972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.103468 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.103899 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:41.603876446 +0000 UTC m=+152.932435260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.198874 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x4zr9"] Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.206428 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.207130 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:41.707102223 +0000 UTC m=+153.035661037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: W1004 04:36:41.252056 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9b027e7_fd36_43e5_80b2_53e105822d71.slice/crio-2b0813fdb72839fc558f5fb48611b83d38defcce8ce02b8fb8c72e9ca9bd8e34 WatchSource:0}: Error finding container 2b0813fdb72839fc558f5fb48611b83d38defcce8ce02b8fb8c72e9ca9bd8e34: Status 404 returned error can't find the container with id 2b0813fdb72839fc558f5fb48611b83d38defcce8ce02b8fb8c72e9ca9bd8e34 Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.308326 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgzgr" event={"ID":"bfad083f-22e8-4845-a327-7457ae07bfb8","Type":"ContainerStarted","Data":"3b5070ab6ddc5414d858275e3f64e9eee44425780fa75e0146989fdfc0b2d953"} Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.308986 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.309307 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:41.80929529 +0000 UTC m=+153.137854094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.354710 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" event={"ID":"eb94e3f3-e813-4122-8718-26d6e005c8f7","Type":"ContainerStarted","Data":"46c59d1dd64017ac900c4c267ee4cb14553198331214a55e55eafd709a8330b0"} Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.402530 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkz66" event={"ID":"7846a3af-bfa1-4129-ae19-48bc580c0a4a","Type":"ContainerStarted","Data":"11a8fd1d569d3442d78d53d1519808fcbff7d8139793924abbd6ac3882101eaa"} Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.409542 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.409878 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:41.909862039 +0000 UTC m=+153.238420853 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.415740 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4zr9" event={"ID":"f9b027e7-fd36-43e5-80b2-53e105822d71","Type":"ContainerStarted","Data":"2b0813fdb72839fc558f5fb48611b83d38defcce8ce02b8fb8c72e9ca9bd8e34"} Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.440203 4575 generic.go:334] "Generic (PLEG): container finished" podID="4b664ad9-7e4d-4330-aa05-620e9617d1c1" containerID="707fb87b72c56fc6f49c3f3f1b8604a6862286c35f647b6e83e2b39ab416c1d1" exitCode=0 Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.440940 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4b664ad9-7e4d-4330-aa05-620e9617d1c1","Type":"ContainerDied","Data":"707fb87b72c56fc6f49c3f3f1b8604a6862286c35f647b6e83e2b39ab416c1d1"} Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.447360 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k5482"] Oct 04 04:36:41 crc kubenswrapper[4575]: W1004 04:36:41.489950 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5b5f157_4684_46a8_bf72_604f9429decc.slice/crio-c75c5b78d8cf39bbc28f0f187723df6f305c7c70eca664831081726444127f63 WatchSource:0}: Error finding container c75c5b78d8cf39bbc28f0f187723df6f305c7c70eca664831081726444127f63: Status 404 returned error can't find the container with id c75c5b78d8cf39bbc28f0f187723df6f305c7c70eca664831081726444127f63 Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.522528 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.523863 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.02384665 +0000 UTC m=+153.352405464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.537517 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rdbvl"] Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.545732 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4caea03c-cd82-4798-b8b1-7007cc09079a" containerName="collect-profiles" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.545767 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4caea03c-cd82-4798-b8b1-7007cc09079a" containerName="collect-profiles" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.545880 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="4caea03c-cd82-4798-b8b1-7007cc09079a" containerName="collect-profiles" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.546751 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.557370 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.559774 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdbvl"] Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.623442 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.623753 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-catalog-content\") pod \"redhat-marketplace-rdbvl\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.623786 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-utilities\") pod \"redhat-marketplace-rdbvl\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.623874 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kdwq\" (UniqueName: \"kubernetes.io/projected/9a908c20-e144-41e6-b331-f5972caf9a33-kube-api-access-7kdwq\") pod \"redhat-marketplace-rdbvl\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.624019 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.123999617 +0000 UTC m=+153.452558441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.726102 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-catalog-content\") pod \"redhat-marketplace-rdbvl\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.726138 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-utilities\") pod \"redhat-marketplace-rdbvl\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.726181 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.726212 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kdwq\" (UniqueName: \"kubernetes.io/projected/9a908c20-e144-41e6-b331-f5972caf9a33-kube-api-access-7kdwq\") pod \"redhat-marketplace-rdbvl\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.726703 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.226683469 +0000 UTC m=+153.555242283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.726860 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-utilities\") pod \"redhat-marketplace-rdbvl\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.727224 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-catalog-content\") pod \"redhat-marketplace-rdbvl\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.731324 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.773185 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kdwq\" (UniqueName: \"kubernetes.io/projected/9a908c20-e144-41e6-b331-f5972caf9a33-kube-api-access-7kdwq\") pod \"redhat-marketplace-rdbvl\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.827783 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.827960 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.327935288 +0000 UTC m=+153.656494102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.828102 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.828440 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.328432773 +0000 UTC m=+153.656991587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.838214 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:41 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:41 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:41 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.838275 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.901871 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.929519 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.929765 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.429717413 +0000 UTC m=+153.758276237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.929827 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:41 crc kubenswrapper[4575]: E1004 04:36:41.930123 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.430111635 +0000 UTC m=+153.758670449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.940550 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnz8"] Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.941817 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:41 crc kubenswrapper[4575]: I1004 04:36:41.956631 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnz8"] Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.035058 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.035270 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-utilities\") pod \"redhat-marketplace-dsnz8\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.035290 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-catalog-content\") pod \"redhat-marketplace-dsnz8\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.035329 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjrvs\" (UniqueName: \"kubernetes.io/projected/7816959d-a0f3-4093-aafa-b16835682a09-kube-api-access-cjrvs\") pod \"redhat-marketplace-dsnz8\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: E1004 04:36:42.035474 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.535444863 +0000 UTC m=+153.864003677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.151029 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.151389 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-utilities\") pod \"redhat-marketplace-dsnz8\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.151417 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-catalog-content\") pod \"redhat-marketplace-dsnz8\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.151468 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjrvs\" (UniqueName: \"kubernetes.io/projected/7816959d-a0f3-4093-aafa-b16835682a09-kube-api-access-cjrvs\") pod \"redhat-marketplace-dsnz8\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: E1004 04:36:42.152120 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.652109432 +0000 UTC m=+153.980668246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.152822 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-utilities\") pod \"redhat-marketplace-dsnz8\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.153106 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-catalog-content\") pod \"redhat-marketplace-dsnz8\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.197845 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjrvs\" (UniqueName: \"kubernetes.io/projected/7816959d-a0f3-4093-aafa-b16835682a09-kube-api-access-cjrvs\") pod \"redhat-marketplace-dsnz8\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.239683 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2q455" Oct 04 04:36:42 crc kubenswrapper[4575]: E1004 04:36:42.254237 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.754206496 +0000 UTC m=+154.082765320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.253934 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.256150 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:42 crc kubenswrapper[4575]: E1004 04:36:42.256562 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.756549414 +0000 UTC m=+154.085108288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.295309 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.344483 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdbvl"] Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.351164 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-66vsx"] Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.352477 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.355657 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.362136 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:42 crc kubenswrapper[4575]: E1004 04:36:42.362800 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.862763848 +0000 UTC m=+154.191322662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.369387 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66vsx"] Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.436171 4575 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.462723 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5b2bec00-534b-4f3e-ab4e-97830d48fb47","Type":"ContainerStarted","Data":"8af08fc13e5a1ecba51a89845216b69df98eceb13862b7ee249b94304eb76c74"} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.462782 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5b2bec00-534b-4f3e-ab4e-97830d48fb47","Type":"ContainerStarted","Data":"1b2366975f337118e7b81c1fb8f35b4652a6f1928798ff2fd9a7a40377ad8a6b"} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.463499 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.463631 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-utilities\") pod \"redhat-operators-66vsx\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.463658 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgcpf\" (UniqueName: \"kubernetes.io/projected/80c6efe3-33be-46c4-b557-ea8fbfed349a-kube-api-access-mgcpf\") pod \"redhat-operators-66vsx\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.463714 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-catalog-content\") pod \"redhat-operators-66vsx\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: E1004 04:36:42.464083 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:36:42.964070699 +0000 UTC m=+154.292629513 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h5dwl" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.465636 4575 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-04T04:36:42.436193147Z","Handler":null,"Name":""} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.471736 4575 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.471770 4575 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.503105 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" event={"ID":"eb94e3f3-e813-4122-8718-26d6e005c8f7","Type":"ContainerStarted","Data":"9c93b3c40f31160e81bdfe9962b8d90427f891e0df106a68276a9c43d1dd082f"} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.513430 4575 generic.go:334] "Generic (PLEG): container finished" podID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerID="59a575ec21f05713d562e47485822d48fa16b53cae915693a0a06a0b93abcd97" exitCode=0 Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.513509 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkz66" event={"ID":"7846a3af-bfa1-4129-ae19-48bc580c0a4a","Type":"ContainerDied","Data":"59a575ec21f05713d562e47485822d48fa16b53cae915693a0a06a0b93abcd97"} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.518741 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.532885 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.532843412 podStartE2EDuration="3.532843412s" podCreationTimestamp="2025-10-04 04:36:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:42.483284429 +0000 UTC m=+153.811843243" watchObservedRunningTime="2025-10-04 04:36:42.532843412 +0000 UTC m=+153.861402226" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.534209 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zljdn"] Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.536867 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.543948 4575 generic.go:334] "Generic (PLEG): container finished" podID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerID="46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b" exitCode=0 Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.544029 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4zr9" event={"ID":"f9b027e7-fd36-43e5-80b2-53e105822d71","Type":"ContainerDied","Data":"46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b"} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.559284 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdbvl" event={"ID":"9a908c20-e144-41e6-b331-f5972caf9a33","Type":"ContainerStarted","Data":"0c9aa185fbde3548114c490edce6591c1504e227893ec467a04283a314f3b3ff"} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.561967 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zljdn"] Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.564670 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.564780 4575 generic.go:334] "Generic (PLEG): container finished" podID="c5b5f157-4684-46a8-bf72-604f9429decc" containerID="475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2" exitCode=0 Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.564832 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5482" event={"ID":"c5b5f157-4684-46a8-bf72-604f9429decc","Type":"ContainerDied","Data":"475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2"} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.564856 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5482" event={"ID":"c5b5f157-4684-46a8-bf72-604f9429decc","Type":"ContainerStarted","Data":"c75c5b78d8cf39bbc28f0f187723df6f305c7c70eca664831081726444127f63"} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.564903 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-utilities\") pod \"redhat-operators-66vsx\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.564938 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgcpf\" (UniqueName: \"kubernetes.io/projected/80c6efe3-33be-46c4-b557-ea8fbfed349a-kube-api-access-mgcpf\") pod \"redhat-operators-66vsx\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.564988 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-catalog-content\") pod \"redhat-operators-66vsx\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.570009 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-catalog-content\") pod \"redhat-operators-66vsx\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.570443 4575 generic.go:334] "Generic (PLEG): container finished" podID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerID="66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce" exitCode=0 Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.571031 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-utilities\") pod \"redhat-operators-66vsx\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.571142 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgzgr" event={"ID":"bfad083f-22e8-4845-a327-7457ae07bfb8","Type":"ContainerDied","Data":"66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce"} Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.588940 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.626794 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgcpf\" (UniqueName: \"kubernetes.io/projected/80c6efe3-33be-46c4-b557-ea8fbfed349a-kube-api-access-mgcpf\") pod \"redhat-operators-66vsx\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.666176 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-catalog-content\") pod \"redhat-operators-zljdn\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.666246 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-utilities\") pod \"redhat-operators-zljdn\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.666356 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.666401 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx5l7\" (UniqueName: \"kubernetes.io/projected/a5ffc246-1302-487b-bcaf-4aee56ed6534-kube-api-access-rx5l7\") pod \"redhat-operators-zljdn\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.705194 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnz8"] Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.706055 4575 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.706073 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.706101 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.756259 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h5dwl\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.767702 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx5l7\" (UniqueName: \"kubernetes.io/projected/a5ffc246-1302-487b-bcaf-4aee56ed6534-kube-api-access-rx5l7\") pod \"redhat-operators-zljdn\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.767749 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-catalog-content\") pod \"redhat-operators-zljdn\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.767775 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-utilities\") pod \"redhat-operators-zljdn\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.768188 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-utilities\") pod \"redhat-operators-zljdn\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.768809 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-catalog-content\") pod \"redhat-operators-zljdn\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.788615 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx5l7\" (UniqueName: \"kubernetes.io/projected/a5ffc246-1302-487b-bcaf-4aee56ed6534-kube-api-access-rx5l7\") pod \"redhat-operators-zljdn\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.838617 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:42 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:42 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:42 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.838671 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.852919 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.868395 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.973704 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kubelet-dir\") pod \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\" (UID: \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\") " Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.973830 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kube-api-access\") pod \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\" (UID: \"4b664ad9-7e4d-4330-aa05-620e9617d1c1\") " Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.973827 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4b664ad9-7e4d-4330-aa05-620e9617d1c1" (UID: "4b664ad9-7e4d-4330-aa05-620e9617d1c1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.974188 4575 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:42 crc kubenswrapper[4575]: I1004 04:36:42.981725 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4b664ad9-7e4d-4330-aa05-620e9617d1c1" (UID: "4b664ad9-7e4d-4330-aa05-620e9617d1c1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.019982 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66vsx"] Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.037506 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.075200 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b664ad9-7e4d-4330-aa05-620e9617d1c1-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.221551 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zljdn"] Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.331756 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.392237 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h5dwl"] Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.578282 4575 generic.go:334] "Generic (PLEG): container finished" podID="5b2bec00-534b-4f3e-ab4e-97830d48fb47" containerID="8af08fc13e5a1ecba51a89845216b69df98eceb13862b7ee249b94304eb76c74" exitCode=0 Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.578415 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5b2bec00-534b-4f3e-ab4e-97830d48fb47","Type":"ContainerDied","Data":"8af08fc13e5a1ecba51a89845216b69df98eceb13862b7ee249b94304eb76c74"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.580651 4575 generic.go:334] "Generic (PLEG): container finished" podID="7816959d-a0f3-4093-aafa-b16835682a09" containerID="54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3" exitCode=0 Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.580704 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnz8" event={"ID":"7816959d-a0f3-4093-aafa-b16835682a09","Type":"ContainerDied","Data":"54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.580725 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnz8" event={"ID":"7816959d-a0f3-4093-aafa-b16835682a09","Type":"ContainerStarted","Data":"68477be674af91e2787e0f8761e54af12490771eb1350904115b9358ba4539ff"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.583665 4575 generic.go:334] "Generic (PLEG): container finished" podID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerID="687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e" exitCode=0 Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.583716 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zljdn" event={"ID":"a5ffc246-1302-487b-bcaf-4aee56ed6534","Type":"ContainerDied","Data":"687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.583733 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zljdn" event={"ID":"a5ffc246-1302-487b-bcaf-4aee56ed6534","Type":"ContainerStarted","Data":"6bb4f7d8b0e8f256d84ed3e66872e93b5a6ea884bbdba6acf8c4c2104d5a5cfc"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.587390 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" event={"ID":"eb94e3f3-e813-4122-8718-26d6e005c8f7","Type":"ContainerStarted","Data":"cca88d884aae2e3d7b38a4fb16cb839822441e5c6f85aac24b54d74dfe9baa9d"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.589384 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4b664ad9-7e4d-4330-aa05-620e9617d1c1","Type":"ContainerDied","Data":"c3de70e669c6d1d94db6ee26d0ea6f323e5b0a3d06ae246de82aa5e4d2147188"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.589416 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3de70e669c6d1d94db6ee26d0ea6f323e5b0a3d06ae246de82aa5e4d2147188" Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.589489 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.590799 4575 generic.go:334] "Generic (PLEG): container finished" podID="9a908c20-e144-41e6-b331-f5972caf9a33" containerID="e10154aa13a9517746a67a2d406e250704f4a38102c0ecc7d1a23eea9fccc90b" exitCode=0 Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.591075 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdbvl" event={"ID":"9a908c20-e144-41e6-b331-f5972caf9a33","Type":"ContainerDied","Data":"e10154aa13a9517746a67a2d406e250704f4a38102c0ecc7d1a23eea9fccc90b"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.592456 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" event={"ID":"2b7b2fb7-8708-45e5-8603-3e7c13f020f8","Type":"ContainerStarted","Data":"60c85b8b74d19a9a0e5f865fc70b7a85331261c42953c850e5144cecd740aeeb"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.596376 4575 generic.go:334] "Generic (PLEG): container finished" podID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerID="e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd" exitCode=0 Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.596422 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vsx" event={"ID":"80c6efe3-33be-46c4-b557-ea8fbfed349a","Type":"ContainerDied","Data":"e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.596469 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vsx" event={"ID":"80c6efe3-33be-46c4-b557-ea8fbfed349a","Type":"ContainerStarted","Data":"e8bfc2936f4ff8db8ec8e402872ce76e82ce4fd6d1c4a905053dc61ffa5ac7f4"} Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.685811 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-lxmpg" podStartSLOduration=16.685794418 podStartE2EDuration="16.685794418s" podCreationTimestamp="2025-10-04 04:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:43.684549891 +0000 UTC m=+155.013108705" watchObservedRunningTime="2025-10-04 04:36:43.685794418 +0000 UTC m=+155.014353222" Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.833654 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:43 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:43 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:43 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:43 crc kubenswrapper[4575]: I1004 04:36:43.833730 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:44 crc kubenswrapper[4575]: I1004 04:36:44.181716 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:44 crc kubenswrapper[4575]: I1004 04:36:44.187494 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-q9ssb" Oct 04 04:36:44 crc kubenswrapper[4575]: I1004 04:36:44.604289 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" event={"ID":"2b7b2fb7-8708-45e5-8603-3e7c13f020f8","Type":"ContainerStarted","Data":"7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7"} Oct 04 04:36:44 crc kubenswrapper[4575]: I1004 04:36:44.845559 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:44 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:44 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:44 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:44 crc kubenswrapper[4575]: I1004 04:36:44.845965 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:44 crc kubenswrapper[4575]: I1004 04:36:44.948538 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.015829 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kube-api-access\") pod \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\" (UID: \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\") " Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.016100 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kubelet-dir\") pod \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\" (UID: \"5b2bec00-534b-4f3e-ab4e-97830d48fb47\") " Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.016173 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5b2bec00-534b-4f3e-ab4e-97830d48fb47" (UID: "5b2bec00-534b-4f3e-ab4e-97830d48fb47"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.016460 4575 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.023076 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5b2bec00-534b-4f3e-ab4e-97830d48fb47" (UID: "5b2bec00-534b-4f3e-ab4e-97830d48fb47"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.117934 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b2bec00-534b-4f3e-ab4e-97830d48fb47-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.585524 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-x6xql" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.617424 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.617432 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5b2bec00-534b-4f3e-ab4e-97830d48fb47","Type":"ContainerDied","Data":"1b2366975f337118e7b81c1fb8f35b4652a6f1928798ff2fd9a7a40377ad8a6b"} Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.617465 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b2366975f337118e7b81c1fb8f35b4652a6f1928798ff2fd9a7a40377ad8a6b" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.617624 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.647560 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" podStartSLOduration=136.647502353 podStartE2EDuration="2m16.647502353s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:45.641693594 +0000 UTC m=+156.970252428" watchObservedRunningTime="2025-10-04 04:36:45.647502353 +0000 UTC m=+156.976061167" Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.831384 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:45 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:45 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:45 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:45 crc kubenswrapper[4575]: I1004 04:36:45.831875 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:46 crc kubenswrapper[4575]: I1004 04:36:46.831266 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:46 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:46 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:46 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:46 crc kubenswrapper[4575]: I1004 04:36:46.831580 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:46 crc kubenswrapper[4575]: I1004 04:36:46.906260 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:36:47 crc kubenswrapper[4575]: I1004 04:36:47.831594 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:47 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:47 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:47 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:47 crc kubenswrapper[4575]: I1004 04:36:47.831666 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:48 crc kubenswrapper[4575]: I1004 04:36:48.830959 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:48 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:48 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:48 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:48 crc kubenswrapper[4575]: I1004 04:36:48.831350 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:49 crc kubenswrapper[4575]: I1004 04:36:49.434669 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:49 crc kubenswrapper[4575]: I1004 04:36:49.434699 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:49 crc kubenswrapper[4575]: I1004 04:36:49.434734 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:49 crc kubenswrapper[4575]: I1004 04:36:49.434735 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:49 crc kubenswrapper[4575]: I1004 04:36:49.643630 4575 patch_prober.go:28] interesting pod/console-f9d7485db-vgclk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.29:8443/health\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 04 04:36:49 crc kubenswrapper[4575]: I1004 04:36:49.643699 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vgclk" podUID="8bf9078b-7040-4b98-86d5-1199c124ccb1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.29:8443/health\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 04 04:36:49 crc kubenswrapper[4575]: I1004 04:36:49.838675 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:49 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:49 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:49 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:49 crc kubenswrapper[4575]: I1004 04:36:49.838750 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:50 crc kubenswrapper[4575]: I1004 04:36:50.226472 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:36:50 crc kubenswrapper[4575]: I1004 04:36:50.830931 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:50 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:50 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:50 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:50 crc kubenswrapper[4575]: I1004 04:36:50.831335 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:51 crc kubenswrapper[4575]: I1004 04:36:51.832189 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:51 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:51 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:51 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:51 crc kubenswrapper[4575]: I1004 04:36:51.832288 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:51 crc kubenswrapper[4575]: I1004 04:36:51.833536 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:51 crc kubenswrapper[4575]: I1004 04:36:51.858239 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/026aaa63-845a-45bb-bc23-7b7c3024eead-metrics-certs\") pod \"network-metrics-daemon-9nvms\" (UID: \"026aaa63-845a-45bb-bc23-7b7c3024eead\") " pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:52 crc kubenswrapper[4575]: I1004 04:36:52.034565 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9nvms" Oct 04 04:36:52 crc kubenswrapper[4575]: I1004 04:36:52.833767 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:52 crc kubenswrapper[4575]: [-]has-synced failed: reason withheld Oct 04 04:36:52 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:52 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:52 crc kubenswrapper[4575]: I1004 04:36:52.833831 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:53 crc kubenswrapper[4575]: I1004 04:36:53.832520 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:36:53 crc kubenswrapper[4575]: [+]has-synced ok Oct 04 04:36:53 crc kubenswrapper[4575]: [+]process-running ok Oct 04 04:36:53 crc kubenswrapper[4575]: healthz check failed Oct 04 04:36:53 crc kubenswrapper[4575]: I1004 04:36:53.832577 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:36:54 crc kubenswrapper[4575]: I1004 04:36:54.832064 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:54 crc kubenswrapper[4575]: I1004 04:36:54.836867 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-bd5sj" Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.434416 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.434781 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.434838 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-9r5mx" Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.434426 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.435162 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.435411 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.435389 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"48a3f37d4be0b8ed7e6877209bb19dc7ef53e56a230ef6e9edcb1a04a5394b42"} pod="openshift-console/downloads-7954f5f757-9r5mx" containerMessage="Container download-server failed liveness probe, will be restarted" Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.435464 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" containerID="cri-o://48a3f37d4be0b8ed7e6877209bb19dc7ef53e56a230ef6e9edcb1a04a5394b42" gracePeriod=2 Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.435433 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.652186 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:36:59 crc kubenswrapper[4575]: I1004 04:36:59.666279 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:37:00 crc kubenswrapper[4575]: I1004 04:37:00.869753 4575 patch_prober.go:28] interesting pod/router-default-5444994796-bd5sj container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 04:37:00 crc kubenswrapper[4575]: I1004 04:37:00.869832 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-bd5sj" podUID="32d2100b-5750-4681-a3c7-5890f03107b7" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:37:01 crc kubenswrapper[4575]: I1004 04:37:01.793099 4575 generic.go:334] "Generic (PLEG): container finished" podID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerID="48a3f37d4be0b8ed7e6877209bb19dc7ef53e56a230ef6e9edcb1a04a5394b42" exitCode=0 Oct 04 04:37:01 crc kubenswrapper[4575]: I1004 04:37:01.793401 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9r5mx" event={"ID":"82bbeaf3-64ae-4676-a365-d75fdf225448","Type":"ContainerDied","Data":"48a3f37d4be0b8ed7e6877209bb19dc7ef53e56a230ef6e9edcb1a04a5394b42"} Oct 04 04:37:03 crc kubenswrapper[4575]: I1004 04:37:03.044026 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:37:08 crc kubenswrapper[4575]: I1004 04:37:08.445977 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:37:08 crc kubenswrapper[4575]: I1004 04:37:08.446324 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:37:09 crc kubenswrapper[4575]: I1004 04:37:09.434424 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:09 crc kubenswrapper[4575]: I1004 04:37:09.434808 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:09 crc kubenswrapper[4575]: I1004 04:37:09.872152 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r5v7h" Oct 04 04:37:17 crc kubenswrapper[4575]: I1004 04:37:17.432241 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:37:19 crc kubenswrapper[4575]: I1004 04:37:19.434119 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:19 crc kubenswrapper[4575]: I1004 04:37:19.434174 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:29 crc kubenswrapper[4575]: I1004 04:37:29.438280 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:29 crc kubenswrapper[4575]: I1004 04:37:29.438852 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:38 crc kubenswrapper[4575]: I1004 04:37:38.446822 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:37:38 crc kubenswrapper[4575]: I1004 04:37:38.448719 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:37:38 crc kubenswrapper[4575]: I1004 04:37:38.448869 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:37:38 crc kubenswrapper[4575]: I1004 04:37:38.449493 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:37:38 crc kubenswrapper[4575]: I1004 04:37:38.449712 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db" gracePeriod=600 Oct 04 04:37:39 crc kubenswrapper[4575]: I1004 04:37:39.434198 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:39 crc kubenswrapper[4575]: I1004 04:37:39.434554 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:41 crc kubenswrapper[4575]: I1004 04:37:41.994319 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db" exitCode=0 Oct 04 04:37:41 crc kubenswrapper[4575]: I1004 04:37:41.994366 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db"} Oct 04 04:37:49 crc kubenswrapper[4575]: I1004 04:37:49.435006 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:49 crc kubenswrapper[4575]: I1004 04:37:49.435630 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:52 crc kubenswrapper[4575]: E1004 04:37:52.708894 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 04:37:52 crc kubenswrapper[4575]: E1004 04:37:52.709070 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nfz6h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-mkz66_openshift-marketplace(7846a3af-bfa1-4129-ae19-48bc580c0a4a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:37:52 crc kubenswrapper[4575]: E1004 04:37:52.710286 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-mkz66" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" Oct 04 04:37:59 crc kubenswrapper[4575]: I1004 04:37:59.434627 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:59 crc kubenswrapper[4575]: I1004 04:37:59.435669 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:38:00 crc kubenswrapper[4575]: E1004 04:38:00.327879 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-mkz66" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" Oct 04 04:38:05 crc kubenswrapper[4575]: E1004 04:38:05.267413 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 04:38:05 crc kubenswrapper[4575]: E1004 04:38:05.267900 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mgcpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-66vsx_openshift-marketplace(80c6efe3-33be-46c4-b557-ea8fbfed349a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:38:05 crc kubenswrapper[4575]: E1004 04:38:05.270754 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-66vsx" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" Oct 04 04:38:09 crc kubenswrapper[4575]: I1004 04:38:09.434196 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:38:09 crc kubenswrapper[4575]: I1004 04:38:09.434524 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:38:19 crc kubenswrapper[4575]: I1004 04:38:19.434292 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:38:19 crc kubenswrapper[4575]: I1004 04:38:19.434837 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.511123 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.511283 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jl6rp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-k5482_openshift-marketplace(c5b5f157-4684-46a8-bf72-604f9429decc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.512391 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-k5482" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.604023 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.604182 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rx5l7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zljdn_openshift-marketplace(a5ffc246-1302-487b-bcaf-4aee56ed6534): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.605354 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zljdn" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.618136 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.618272 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vx4jd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dgzgr_openshift-marketplace(bfad083f-22e8-4845-a327-7457ae07bfb8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.619395 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dgzgr" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.710092 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.710263 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-84dd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-x4zr9_openshift-marketplace(f9b027e7-fd36-43e5-80b2-53e105822d71): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:38:19 crc kubenswrapper[4575]: E1004 04:38:19.711369 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-x4zr9" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" Oct 04 04:38:22 crc kubenswrapper[4575]: E1004 04:38:22.316141 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dgzgr" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" Oct 04 04:38:22 crc kubenswrapper[4575]: E1004 04:38:22.316141 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-k5482" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" Oct 04 04:38:22 crc kubenswrapper[4575]: E1004 04:38:22.316373 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-x4zr9" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" Oct 04 04:38:22 crc kubenswrapper[4575]: E1004 04:38:22.316408 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zljdn" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" Oct 04 04:38:24 crc kubenswrapper[4575]: E1004 04:38:24.752008 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 04:38:24 crc kubenswrapper[4575]: E1004 04:38:24.752420 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7kdwq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rdbvl_openshift-marketplace(9a908c20-e144-41e6-b331-f5972caf9a33): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:38:24 crc kubenswrapper[4575]: E1004 04:38:24.753710 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rdbvl" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" Oct 04 04:38:25 crc kubenswrapper[4575]: I1004 04:38:25.084946 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9nvms"] Oct 04 04:38:25 crc kubenswrapper[4575]: W1004 04:38:25.091324 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod026aaa63_845a_45bb_bc23_7b7c3024eead.slice/crio-9402c8e9018a2f4150e53eac17561880d4b431487e8aa18f8a680b47e4b8fdb2 WatchSource:0}: Error finding container 9402c8e9018a2f4150e53eac17561880d4b431487e8aa18f8a680b47e4b8fdb2: Status 404 returned error can't find the container with id 9402c8e9018a2f4150e53eac17561880d4b431487e8aa18f8a680b47e4b8fdb2 Oct 04 04:38:25 crc kubenswrapper[4575]: I1004 04:38:25.194422 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9nvms" event={"ID":"026aaa63-845a-45bb-bc23-7b7c3024eead","Type":"ContainerStarted","Data":"9402c8e9018a2f4150e53eac17561880d4b431487e8aa18f8a680b47e4b8fdb2"} Oct 04 04:38:25 crc kubenswrapper[4575]: E1004 04:38:25.195755 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rdbvl" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" Oct 04 04:38:25 crc kubenswrapper[4575]: E1004 04:38:25.649490 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 04:38:25 crc kubenswrapper[4575]: E1004 04:38:25.649896 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cjrvs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dsnz8_openshift-marketplace(7816959d-a0f3-4093-aafa-b16835682a09): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:38:25 crc kubenswrapper[4575]: E1004 04:38:25.651075 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dsnz8" podUID="7816959d-a0f3-4093-aafa-b16835682a09" Oct 04 04:38:26 crc kubenswrapper[4575]: I1004 04:38:26.201861 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"7df48bb006932f578e73ad68f06d44012c4eb713d15727b98e4c310b3bde01ac"} Oct 04 04:38:26 crc kubenswrapper[4575]: I1004 04:38:26.203530 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9r5mx" event={"ID":"82bbeaf3-64ae-4676-a365-d75fdf225448","Type":"ContainerStarted","Data":"f7915528f05b5328fb7bfb470798e31df0c9db52a2dd30afa1626e1376d93af2"} Oct 04 04:38:26 crc kubenswrapper[4575]: I1004 04:38:26.203959 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9r5mx" Oct 04 04:38:26 crc kubenswrapper[4575]: I1004 04:38:26.204076 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:38:26 crc kubenswrapper[4575]: I1004 04:38:26.204114 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:38:26 crc kubenswrapper[4575]: I1004 04:38:26.205312 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9nvms" event={"ID":"026aaa63-845a-45bb-bc23-7b7c3024eead","Type":"ContainerStarted","Data":"8217e11d8597142668b76f51e25ba0a2db77d533b7502de5265a1d6132278229"} Oct 04 04:38:26 crc kubenswrapper[4575]: I1004 04:38:26.205347 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9nvms" event={"ID":"026aaa63-845a-45bb-bc23-7b7c3024eead","Type":"ContainerStarted","Data":"8b14e7e6a10313ffb8fe04fccf97b89cf9f85edd6079b8ae2fb251d5bc5acb1f"} Oct 04 04:38:26 crc kubenswrapper[4575]: E1004 04:38:26.206800 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dsnz8" podUID="7816959d-a0f3-4093-aafa-b16835682a09" Oct 04 04:38:26 crc kubenswrapper[4575]: I1004 04:38:26.248549 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9nvms" podStartSLOduration=237.248533968 podStartE2EDuration="3m57.248533968s" podCreationTimestamp="2025-10-04 04:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:38:26.246782106 +0000 UTC m=+257.575340930" watchObservedRunningTime="2025-10-04 04:38:26.248533968 +0000 UTC m=+257.577092782" Oct 04 04:38:27 crc kubenswrapper[4575]: I1004 04:38:27.211405 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:38:27 crc kubenswrapper[4575]: I1004 04:38:27.211450 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:38:29 crc kubenswrapper[4575]: I1004 04:38:29.434893 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:38:29 crc kubenswrapper[4575]: I1004 04:38:29.435227 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:38:29 crc kubenswrapper[4575]: I1004 04:38:29.434904 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:38:29 crc kubenswrapper[4575]: I1004 04:38:29.435358 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:38:39 crc kubenswrapper[4575]: I1004 04:38:39.435056 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:38:39 crc kubenswrapper[4575]: I1004 04:38:39.435255 4575 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5mx container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:38:39 crc kubenswrapper[4575]: I1004 04:38:39.435477 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:38:39 crc kubenswrapper[4575]: I1004 04:38:39.435399 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5mx" podUID="82bbeaf3-64ae-4676-a365-d75fdf225448" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:38:44 crc kubenswrapper[4575]: I1004 04:38:44.290221 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkz66" event={"ID":"7846a3af-bfa1-4129-ae19-48bc580c0a4a","Type":"ContainerStarted","Data":"a974e424ae7b55d181273802ef21b632423d4a0f42e9db87601e63dc42330403"} Oct 04 04:38:44 crc kubenswrapper[4575]: I1004 04:38:44.292380 4575 generic.go:334] "Generic (PLEG): container finished" podID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerID="2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49" exitCode=0 Oct 04 04:38:44 crc kubenswrapper[4575]: I1004 04:38:44.292416 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vsx" event={"ID":"80c6efe3-33be-46c4-b557-ea8fbfed349a","Type":"ContainerDied","Data":"2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49"} Oct 04 04:38:45 crc kubenswrapper[4575]: I1004 04:38:45.299662 4575 generic.go:334] "Generic (PLEG): container finished" podID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerID="a974e424ae7b55d181273802ef21b632423d4a0f42e9db87601e63dc42330403" exitCode=0 Oct 04 04:38:45 crc kubenswrapper[4575]: I1004 04:38:45.299710 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkz66" event={"ID":"7846a3af-bfa1-4129-ae19-48bc580c0a4a","Type":"ContainerDied","Data":"a974e424ae7b55d181273802ef21b632423d4a0f42e9db87601e63dc42330403"} Oct 04 04:38:49 crc kubenswrapper[4575]: I1004 04:38:49.456018 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-9r5mx" Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.411416 4575 generic.go:334] "Generic (PLEG): container finished" podID="9a908c20-e144-41e6-b331-f5972caf9a33" containerID="8a7edd9d85fbde7cdefd4cad1fafb6f0b9b21c71d7780f6c8f88cdf91a780aab" exitCode=0 Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.412154 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdbvl" event={"ID":"9a908c20-e144-41e6-b331-f5972caf9a33","Type":"ContainerDied","Data":"8a7edd9d85fbde7cdefd4cad1fafb6f0b9b21c71d7780f6c8f88cdf91a780aab"} Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.414846 4575 generic.go:334] "Generic (PLEG): container finished" podID="c5b5f157-4684-46a8-bf72-604f9429decc" containerID="8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec" exitCode=0 Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.414893 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5482" event={"ID":"c5b5f157-4684-46a8-bf72-604f9429decc","Type":"ContainerDied","Data":"8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec"} Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.436943 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vsx" event={"ID":"80c6efe3-33be-46c4-b557-ea8fbfed349a","Type":"ContainerStarted","Data":"16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab"} Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.441728 4575 generic.go:334] "Generic (PLEG): container finished" podID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerID="3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9" exitCode=0 Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.441812 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgzgr" event={"ID":"bfad083f-22e8-4845-a327-7457ae07bfb8","Type":"ContainerDied","Data":"3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9"} Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.445296 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnz8" event={"ID":"7816959d-a0f3-4093-aafa-b16835682a09","Type":"ContainerDied","Data":"0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb"} Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.445218 4575 generic.go:334] "Generic (PLEG): container finished" podID="7816959d-a0f3-4093-aafa-b16835682a09" containerID="0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb" exitCode=0 Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.460280 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkz66" event={"ID":"7846a3af-bfa1-4129-ae19-48bc580c0a4a","Type":"ContainerStarted","Data":"4c6a7d047e294e4598b55c4686e14eef2aa50b089cb2a3e926c6b8f21eb738a7"} Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.462407 4575 generic.go:334] "Generic (PLEG): container finished" podID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerID="72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78" exitCode=0 Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.462455 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zljdn" event={"ID":"a5ffc246-1302-487b-bcaf-4aee56ed6534","Type":"ContainerDied","Data":"72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78"} Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.470212 4575 generic.go:334] "Generic (PLEG): container finished" podID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerID="4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26" exitCode=0 Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.470257 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4zr9" event={"ID":"f9b027e7-fd36-43e5-80b2-53e105822d71","Type":"ContainerDied","Data":"4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26"} Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.471389 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-66vsx" podStartSLOduration=2.623162512 podStartE2EDuration="2m21.471376308s" podCreationTimestamp="2025-10-04 04:36:42 +0000 UTC" firstStartedPulling="2025-10-04 04:36:43.598150714 +0000 UTC m=+154.926709528" lastFinishedPulling="2025-10-04 04:39:02.44636449 +0000 UTC m=+293.774923324" observedRunningTime="2025-10-04 04:39:03.469485972 +0000 UTC m=+294.798044786" watchObservedRunningTime="2025-10-04 04:39:03.471376308 +0000 UTC m=+294.799935122" Oct 04 04:39:03 crc kubenswrapper[4575]: I1004 04:39:03.557752 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mkz66" podStartSLOduration=5.444021235 podStartE2EDuration="2m25.557735486s" podCreationTimestamp="2025-10-04 04:36:38 +0000 UTC" firstStartedPulling="2025-10-04 04:36:42.518413712 +0000 UTC m=+153.846972526" lastFinishedPulling="2025-10-04 04:39:02.632127963 +0000 UTC m=+293.960686777" observedRunningTime="2025-10-04 04:39:03.55650097 +0000 UTC m=+294.885059794" watchObservedRunningTime="2025-10-04 04:39:03.557735486 +0000 UTC m=+294.886294300" Oct 04 04:39:09 crc kubenswrapper[4575]: I1004 04:39:09.101978 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:39:09 crc kubenswrapper[4575]: I1004 04:39:09.102530 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:39:10 crc kubenswrapper[4575]: I1004 04:39:10.173982 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:39:10 crc kubenswrapper[4575]: I1004 04:39:10.230613 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:39:12 crc kubenswrapper[4575]: I1004 04:39:12.706929 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:39:12 crc kubenswrapper[4575]: I1004 04:39:12.707279 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:39:12 crc kubenswrapper[4575]: I1004 04:39:12.754258 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:39:13 crc kubenswrapper[4575]: I1004 04:39:13.562489 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.559024 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zljdn" event={"ID":"a5ffc246-1302-487b-bcaf-4aee56ed6534","Type":"ContainerStarted","Data":"af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216"} Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.562771 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4zr9" event={"ID":"f9b027e7-fd36-43e5-80b2-53e105822d71","Type":"ContainerStarted","Data":"f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3"} Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.565261 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdbvl" event={"ID":"9a908c20-e144-41e6-b331-f5972caf9a33","Type":"ContainerStarted","Data":"a086d332af4c29416e9fefa267029dea60413a85427e03abb322dc592e23980f"} Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.567607 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5482" event={"ID":"c5b5f157-4684-46a8-bf72-604f9429decc","Type":"ContainerStarted","Data":"fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658"} Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.570426 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgzgr" event={"ID":"bfad083f-22e8-4845-a327-7457ae07bfb8","Type":"ContainerStarted","Data":"368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4"} Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.572910 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnz8" event={"ID":"7816959d-a0f3-4093-aafa-b16835682a09","Type":"ContainerStarted","Data":"9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f"} Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.603065 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k5482" podStartSLOduration=4.620248299 podStartE2EDuration="2m41.603047389s" podCreationTimestamp="2025-10-04 04:36:39 +0000 UTC" firstStartedPulling="2025-10-04 04:36:42.571656523 +0000 UTC m=+153.900215337" lastFinishedPulling="2025-10-04 04:39:19.554455623 +0000 UTC m=+310.883014427" observedRunningTime="2025-10-04 04:39:20.599502774 +0000 UTC m=+311.928061598" watchObservedRunningTime="2025-10-04 04:39:20.603047389 +0000 UTC m=+311.931606203" Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.604882 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zljdn" podStartSLOduration=3.67420968 podStartE2EDuration="2m38.604873423s" podCreationTimestamp="2025-10-04 04:36:42 +0000 UTC" firstStartedPulling="2025-10-04 04:36:44.606966002 +0000 UTC m=+155.935524816" lastFinishedPulling="2025-10-04 04:39:19.537629745 +0000 UTC m=+310.866188559" observedRunningTime="2025-10-04 04:39:20.583040767 +0000 UTC m=+311.911599601" watchObservedRunningTime="2025-10-04 04:39:20.604873423 +0000 UTC m=+311.933432237" Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.620309 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rdbvl" podStartSLOduration=9.729333624 podStartE2EDuration="2m39.62028833s" podCreationTimestamp="2025-10-04 04:36:41 +0000 UTC" firstStartedPulling="2025-10-04 04:36:43.592318155 +0000 UTC m=+154.920876969" lastFinishedPulling="2025-10-04 04:39:13.483272861 +0000 UTC m=+304.811831675" observedRunningTime="2025-10-04 04:39:20.61790657 +0000 UTC m=+311.946465404" watchObservedRunningTime="2025-10-04 04:39:20.62028833 +0000 UTC m=+311.948847144" Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.646256 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dsnz8" podStartSLOduration=3.754593983 podStartE2EDuration="2m39.646239909s" podCreationTimestamp="2025-10-04 04:36:41 +0000 UTC" firstStartedPulling="2025-10-04 04:36:43.582118757 +0000 UTC m=+154.910677571" lastFinishedPulling="2025-10-04 04:39:19.473764683 +0000 UTC m=+310.802323497" observedRunningTime="2025-10-04 04:39:20.644828717 +0000 UTC m=+311.973387541" watchObservedRunningTime="2025-10-04 04:39:20.646239909 +0000 UTC m=+311.974798723" Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.666685 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dgzgr" podStartSLOduration=4.737462843 podStartE2EDuration="2m41.666671614s" podCreationTimestamp="2025-10-04 04:36:39 +0000 UTC" firstStartedPulling="2025-10-04 04:36:42.608532817 +0000 UTC m=+153.937091631" lastFinishedPulling="2025-10-04 04:39:19.537741588 +0000 UTC m=+310.866300402" observedRunningTime="2025-10-04 04:39:20.664554132 +0000 UTC m=+311.993112956" watchObservedRunningTime="2025-10-04 04:39:20.666671614 +0000 UTC m=+311.995230428" Oct 04 04:39:20 crc kubenswrapper[4575]: I1004 04:39:20.684806 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x4zr9" podStartSLOduration=4.694841991 podStartE2EDuration="2m41.684786121s" podCreationTimestamp="2025-10-04 04:36:39 +0000 UTC" firstStartedPulling="2025-10-04 04:36:42.547650584 +0000 UTC m=+153.876209398" lastFinishedPulling="2025-10-04 04:39:19.537594714 +0000 UTC m=+310.866153528" observedRunningTime="2025-10-04 04:39:20.681562355 +0000 UTC m=+312.010121169" watchObservedRunningTime="2025-10-04 04:39:20.684786121 +0000 UTC m=+312.013344935" Oct 04 04:39:21 crc kubenswrapper[4575]: I1004 04:39:21.905768 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:39:21 crc kubenswrapper[4575]: I1004 04:39:21.906107 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:39:21 crc kubenswrapper[4575]: I1004 04:39:21.959356 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:39:22 crc kubenswrapper[4575]: I1004 04:39:22.296212 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:39:22 crc kubenswrapper[4575]: I1004 04:39:22.296509 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:39:22 crc kubenswrapper[4575]: I1004 04:39:22.336063 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:39:22 crc kubenswrapper[4575]: I1004 04:39:22.869330 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:39:22 crc kubenswrapper[4575]: I1004 04:39:22.869389 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:39:23 crc kubenswrapper[4575]: I1004 04:39:23.905165 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zljdn" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerName="registry-server" probeResult="failure" output=< Oct 04 04:39:23 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 04:39:23 crc kubenswrapper[4575]: > Oct 04 04:39:29 crc kubenswrapper[4575]: I1004 04:39:29.504025 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:39:29 crc kubenswrapper[4575]: I1004 04:39:29.504293 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:39:29 crc kubenswrapper[4575]: I1004 04:39:29.543385 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:39:29 crc kubenswrapper[4575]: I1004 04:39:29.651909 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:39:29 crc kubenswrapper[4575]: I1004 04:39:29.770778 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dgzgr"] Oct 04 04:39:29 crc kubenswrapper[4575]: I1004 04:39:29.893550 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:39:29 crc kubenswrapper[4575]: I1004 04:39:29.893623 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:39:29 crc kubenswrapper[4575]: I1004 04:39:29.935200 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:39:30 crc kubenswrapper[4575]: I1004 04:39:30.226912 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k5482" Oct 04 04:39:30 crc kubenswrapper[4575]: I1004 04:39:30.226957 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k5482" Oct 04 04:39:30 crc kubenswrapper[4575]: I1004 04:39:30.293407 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k5482" Oct 04 04:39:30 crc kubenswrapper[4575]: I1004 04:39:30.658700 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k5482" Oct 04 04:39:30 crc kubenswrapper[4575]: I1004 04:39:30.666830 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:39:31 crc kubenswrapper[4575]: I1004 04:39:31.624684 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dgzgr" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerName="registry-server" containerID="cri-o://368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4" gracePeriod=2 Oct 04 04:39:31 crc kubenswrapper[4575]: I1004 04:39:31.953972 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:39:31 crc kubenswrapper[4575]: I1004 04:39:31.965193 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.117639 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-utilities\") pod \"bfad083f-22e8-4845-a327-7457ae07bfb8\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.117769 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-catalog-content\") pod \"bfad083f-22e8-4845-a327-7457ae07bfb8\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.118778 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-utilities" (OuterVolumeSpecName: "utilities") pod "bfad083f-22e8-4845-a327-7457ae07bfb8" (UID: "bfad083f-22e8-4845-a327-7457ae07bfb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.125764 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfad083f-22e8-4845-a327-7457ae07bfb8-kube-api-access-vx4jd" (OuterVolumeSpecName: "kube-api-access-vx4jd") pod "bfad083f-22e8-4845-a327-7457ae07bfb8" (UID: "bfad083f-22e8-4845-a327-7457ae07bfb8"). InnerVolumeSpecName "kube-api-access-vx4jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.117804 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx4jd\" (UniqueName: \"kubernetes.io/projected/bfad083f-22e8-4845-a327-7457ae07bfb8-kube-api-access-vx4jd\") pod \"bfad083f-22e8-4845-a327-7457ae07bfb8\" (UID: \"bfad083f-22e8-4845-a327-7457ae07bfb8\") " Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.126294 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.126320 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx4jd\" (UniqueName: \"kubernetes.io/projected/bfad083f-22e8-4845-a327-7457ae07bfb8-kube-api-access-vx4jd\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.161343 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfad083f-22e8-4845-a327-7457ae07bfb8" (UID: "bfad083f-22e8-4845-a327-7457ae07bfb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.171688 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k5482"] Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.227272 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfad083f-22e8-4845-a327-7457ae07bfb8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.334709 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.632360 4575 generic.go:334] "Generic (PLEG): container finished" podID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerID="368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4" exitCode=0 Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.633124 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgzgr" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.640913 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgzgr" event={"ID":"bfad083f-22e8-4845-a327-7457ae07bfb8","Type":"ContainerDied","Data":"368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4"} Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.640980 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgzgr" event={"ID":"bfad083f-22e8-4845-a327-7457ae07bfb8","Type":"ContainerDied","Data":"3b5070ab6ddc5414d858275e3f64e9eee44425780fa75e0146989fdfc0b2d953"} Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.641000 4575 scope.go:117] "RemoveContainer" containerID="368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.641076 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k5482" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" containerName="registry-server" containerID="cri-o://fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658" gracePeriod=2 Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.669483 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dgzgr"] Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.673529 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dgzgr"] Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.673721 4575 scope.go:117] "RemoveContainer" containerID="3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9" Oct 04 04:39:32 crc kubenswrapper[4575]: E1004 04:39:32.703523 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfad083f_22e8_4845_a327_7457ae07bfb8.slice/crio-3b5070ab6ddc5414d858275e3f64e9eee44425780fa75e0146989fdfc0b2d953\": RecentStats: unable to find data in memory cache]" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.705509 4575 scope.go:117] "RemoveContainer" containerID="66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.751444 4575 scope.go:117] "RemoveContainer" containerID="368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4" Oct 04 04:39:32 crc kubenswrapper[4575]: E1004 04:39:32.754827 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4\": container with ID starting with 368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4 not found: ID does not exist" containerID="368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.754873 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4"} err="failed to get container status \"368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4\": rpc error: code = NotFound desc = could not find container \"368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4\": container with ID starting with 368a3bdc0d8aaeee2740c05f8e88462baa82f0eee88dfaec624768f9a7ea9cc4 not found: ID does not exist" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.754902 4575 scope.go:117] "RemoveContainer" containerID="3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9" Oct 04 04:39:32 crc kubenswrapper[4575]: E1004 04:39:32.755402 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9\": container with ID starting with 3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9 not found: ID does not exist" containerID="3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.755422 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9"} err="failed to get container status \"3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9\": rpc error: code = NotFound desc = could not find container \"3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9\": container with ID starting with 3bb8f7ce7868c709a9a8236930731039e67a63ac4b36bb9a7485eca1ebf4dca9 not found: ID does not exist" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.755448 4575 scope.go:117] "RemoveContainer" containerID="66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce" Oct 04 04:39:32 crc kubenswrapper[4575]: E1004 04:39:32.755750 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce\": container with ID starting with 66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce not found: ID does not exist" containerID="66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.755788 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce"} err="failed to get container status \"66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce\": rpc error: code = NotFound desc = could not find container \"66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce\": container with ID starting with 66a4987c6035cb414d9fb64c96e382198f3e09b0d39dd4664f3ada85526740ce not found: ID does not exist" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.905436 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:39:32 crc kubenswrapper[4575]: I1004 04:39:32.952122 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.316025 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" path="/var/lib/kubelet/pods/bfad083f-22e8-4845-a327-7457ae07bfb8/volumes" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.493578 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k5482" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.640973 4575 generic.go:334] "Generic (PLEG): container finished" podID="c5b5f157-4684-46a8-bf72-604f9429decc" containerID="fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658" exitCode=0 Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.641048 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k5482" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.641069 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5482" event={"ID":"c5b5f157-4684-46a8-bf72-604f9429decc","Type":"ContainerDied","Data":"fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658"} Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.641185 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5482" event={"ID":"c5b5f157-4684-46a8-bf72-604f9429decc","Type":"ContainerDied","Data":"c75c5b78d8cf39bbc28f0f187723df6f305c7c70eca664831081726444127f63"} Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.641219 4575 scope.go:117] "RemoveContainer" containerID="fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.648128 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl6rp\" (UniqueName: \"kubernetes.io/projected/c5b5f157-4684-46a8-bf72-604f9429decc-kube-api-access-jl6rp\") pod \"c5b5f157-4684-46a8-bf72-604f9429decc\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.648304 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-utilities\") pod \"c5b5f157-4684-46a8-bf72-604f9429decc\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.648344 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-catalog-content\") pod \"c5b5f157-4684-46a8-bf72-604f9429decc\" (UID: \"c5b5f157-4684-46a8-bf72-604f9429decc\") " Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.649163 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-utilities" (OuterVolumeSpecName: "utilities") pod "c5b5f157-4684-46a8-bf72-604f9429decc" (UID: "c5b5f157-4684-46a8-bf72-604f9429decc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.660843 4575 scope.go:117] "RemoveContainer" containerID="8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.660946 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5b5f157-4684-46a8-bf72-604f9429decc-kube-api-access-jl6rp" (OuterVolumeSpecName: "kube-api-access-jl6rp") pod "c5b5f157-4684-46a8-bf72-604f9429decc" (UID: "c5b5f157-4684-46a8-bf72-604f9429decc"). InnerVolumeSpecName "kube-api-access-jl6rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.683065 4575 scope.go:117] "RemoveContainer" containerID="475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.704210 4575 scope.go:117] "RemoveContainer" containerID="fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.704653 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658\": container with ID starting with fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658 not found: ID does not exist" containerID="fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.704680 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658"} err="failed to get container status \"fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658\": rpc error: code = NotFound desc = could not find container \"fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658\": container with ID starting with fe01ac2d58ff7a390c354c9edc718e58f95b7a090188fe74c8139e579b8bb658 not found: ID does not exist" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.704700 4575 scope.go:117] "RemoveContainer" containerID="8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.705086 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec\": container with ID starting with 8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec not found: ID does not exist" containerID="8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.705116 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec"} err="failed to get container status \"8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec\": rpc error: code = NotFound desc = could not find container \"8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec\": container with ID starting with 8e845bf4469713ee77f82b2a117451c065cb323fcb1197eb9922b46ea5e464ec not found: ID does not exist" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.705128 4575 scope.go:117] "RemoveContainer" containerID="475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.705862 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2\": container with ID starting with 475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2 not found: ID does not exist" containerID="475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.705906 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2"} err="failed to get container status \"475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2\": rpc error: code = NotFound desc = could not find container \"475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2\": container with ID starting with 475162af752c570d5cfa0f9f80e9ca6ab33152b88a7602c2a4ec90613f366dd2 not found: ID does not exist" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.716886 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5b5f157-4684-46a8-bf72-604f9429decc" (UID: "c5b5f157-4684-46a8-bf72-604f9429decc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.749536 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.749595 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl6rp\" (UniqueName: \"kubernetes.io/projected/c5b5f157-4684-46a8-bf72-604f9429decc-kube-api-access-jl6rp\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.749607 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5b5f157-4684-46a8-bf72-604f9429decc-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.774748 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bmzcj"] Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.774939 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerName="extract-content" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.774949 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerName="extract-content" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.774960 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerName="registry-server" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.774966 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerName="registry-server" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.774973 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" containerName="registry-server" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.774979 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" containerName="registry-server" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.774992 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerName="extract-utilities" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.774998 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerName="extract-utilities" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.775008 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b664ad9-7e4d-4330-aa05-620e9617d1c1" containerName="pruner" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.775016 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b664ad9-7e4d-4330-aa05-620e9617d1c1" containerName="pruner" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.775024 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" containerName="extract-utilities" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.775031 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" containerName="extract-utilities" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.775041 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b2bec00-534b-4f3e-ab4e-97830d48fb47" containerName="pruner" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.775046 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b2bec00-534b-4f3e-ab4e-97830d48fb47" containerName="pruner" Oct 04 04:39:33 crc kubenswrapper[4575]: E1004 04:39:33.775054 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" containerName="extract-content" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.775061 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" containerName="extract-content" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.775150 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" containerName="registry-server" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.775161 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b2bec00-534b-4f3e-ab4e-97830d48fb47" containerName="pruner" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.775172 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b664ad9-7e4d-4330-aa05-620e9617d1c1" containerName="pruner" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.775181 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfad083f-22e8-4845-a327-7457ae07bfb8" containerName="registry-server" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.775524 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.826809 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bmzcj"] Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.952254 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.952323 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-registry-certificates\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.952354 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.952380 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-bound-sa-token\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.952410 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-registry-tls\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.952433 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.952481 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97fj7\" (UniqueName: \"kubernetes.io/projected/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-kube-api-access-97fj7\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.952525 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-trusted-ca\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.967148 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k5482"] Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.971022 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k5482"] Oct 04 04:39:33 crc kubenswrapper[4575]: I1004 04:39:33.985724 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.054206 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-bound-sa-token\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.054300 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-registry-tls\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.054339 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.054366 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97fj7\" (UniqueName: \"kubernetes.io/projected/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-kube-api-access-97fj7\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.054398 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-trusted-ca\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.054444 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-registry-certificates\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.054500 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.055077 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.056174 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-trusted-ca\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.056279 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-registry-certificates\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.058887 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.059028 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-registry-tls\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.072362 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97fj7\" (UniqueName: \"kubernetes.io/projected/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-kube-api-access-97fj7\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.075917 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/68b00b8a-c992-4bcb-84b8-a9d03e239e0e-bound-sa-token\") pod \"image-registry-66df7c8f76-bmzcj\" (UID: \"68b00b8a-c992-4bcb-84b8-a9d03e239e0e\") " pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.094380 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.286024 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bmzcj"] Oct 04 04:39:34 crc kubenswrapper[4575]: W1004 04:39:34.296059 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68b00b8a_c992_4bcb_84b8_a9d03e239e0e.slice/crio-cf88be856f4345ec240e77dea50f5525d60321d66f3b2bb6246b0d0069f05f37 WatchSource:0}: Error finding container cf88be856f4345ec240e77dea50f5525d60321d66f3b2bb6246b0d0069f05f37: Status 404 returned error can't find the container with id cf88be856f4345ec240e77dea50f5525d60321d66f3b2bb6246b0d0069f05f37 Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.374924 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnz8"] Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.375600 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dsnz8" podUID="7816959d-a0f3-4093-aafa-b16835682a09" containerName="registry-server" containerID="cri-o://9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f" gracePeriod=2 Oct 04 04:39:34 crc kubenswrapper[4575]: I1004 04:39:34.648572 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" event={"ID":"68b00b8a-c992-4bcb-84b8-a9d03e239e0e","Type":"ContainerStarted","Data":"cf88be856f4345ec240e77dea50f5525d60321d66f3b2bb6246b0d0069f05f37"} Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.207535 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.321844 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5b5f157-4684-46a8-bf72-604f9429decc" path="/var/lib/kubelet/pods/c5b5f157-4684-46a8-bf72-604f9429decc/volumes" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.371141 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-catalog-content\") pod \"7816959d-a0f3-4093-aafa-b16835682a09\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.371233 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjrvs\" (UniqueName: \"kubernetes.io/projected/7816959d-a0f3-4093-aafa-b16835682a09-kube-api-access-cjrvs\") pod \"7816959d-a0f3-4093-aafa-b16835682a09\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.371344 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-utilities\") pod \"7816959d-a0f3-4093-aafa-b16835682a09\" (UID: \"7816959d-a0f3-4093-aafa-b16835682a09\") " Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.372316 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-utilities" (OuterVolumeSpecName: "utilities") pod "7816959d-a0f3-4093-aafa-b16835682a09" (UID: "7816959d-a0f3-4093-aafa-b16835682a09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.372736 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.376924 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7816959d-a0f3-4093-aafa-b16835682a09-kube-api-access-cjrvs" (OuterVolumeSpecName: "kube-api-access-cjrvs") pod "7816959d-a0f3-4093-aafa-b16835682a09" (UID: "7816959d-a0f3-4093-aafa-b16835682a09"). InnerVolumeSpecName "kube-api-access-cjrvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.383720 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7816959d-a0f3-4093-aafa-b16835682a09" (UID: "7816959d-a0f3-4093-aafa-b16835682a09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.474485 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjrvs\" (UniqueName: \"kubernetes.io/projected/7816959d-a0f3-4093-aafa-b16835682a09-kube-api-access-cjrvs\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.474530 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7816959d-a0f3-4093-aafa-b16835682a09-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.655251 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" event={"ID":"68b00b8a-c992-4bcb-84b8-a9d03e239e0e","Type":"ContainerStarted","Data":"ec801fe6906b383048bbf2d09329552890f9952a5fd81a356f291fd1e9d35c40"} Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.655413 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.657447 4575 generic.go:334] "Generic (PLEG): container finished" podID="7816959d-a0f3-4093-aafa-b16835682a09" containerID="9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f" exitCode=0 Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.657510 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnz8" event={"ID":"7816959d-a0f3-4093-aafa-b16835682a09","Type":"ContainerDied","Data":"9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f"} Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.657552 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsnz8" event={"ID":"7816959d-a0f3-4093-aafa-b16835682a09","Type":"ContainerDied","Data":"68477be674af91e2787e0f8761e54af12490771eb1350904115b9358ba4539ff"} Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.657557 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsnz8" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.657576 4575 scope.go:117] "RemoveContainer" containerID="9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.683597 4575 scope.go:117] "RemoveContainer" containerID="0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.703005 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" podStartSLOduration=2.702988813 podStartE2EDuration="2.702988813s" podCreationTimestamp="2025-10-04 04:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:39:35.681607769 +0000 UTC m=+327.010166593" watchObservedRunningTime="2025-10-04 04:39:35.702988813 +0000 UTC m=+327.031547637" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.711355 4575 scope.go:117] "RemoveContainer" containerID="54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.713048 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnz8"] Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.713473 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsnz8"] Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.727055 4575 scope.go:117] "RemoveContainer" containerID="9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f" Oct 04 04:39:35 crc kubenswrapper[4575]: E1004 04:39:35.727436 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f\": container with ID starting with 9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f not found: ID does not exist" containerID="9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.727470 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f"} err="failed to get container status \"9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f\": rpc error: code = NotFound desc = could not find container \"9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f\": container with ID starting with 9826cd7539500f1160b4e87b78e0583bd340405662d604df6ed45aa42e1e614f not found: ID does not exist" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.727495 4575 scope.go:117] "RemoveContainer" containerID="0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb" Oct 04 04:39:35 crc kubenswrapper[4575]: E1004 04:39:35.727881 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb\": container with ID starting with 0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb not found: ID does not exist" containerID="0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.727905 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb"} err="failed to get container status \"0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb\": rpc error: code = NotFound desc = could not find container \"0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb\": container with ID starting with 0dd0c33f01aee554df254c07311e4a307aafa1aee200ffc26f0d93fcac1c79bb not found: ID does not exist" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.727925 4575 scope.go:117] "RemoveContainer" containerID="54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3" Oct 04 04:39:35 crc kubenswrapper[4575]: E1004 04:39:35.728113 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3\": container with ID starting with 54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3 not found: ID does not exist" containerID="54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3" Oct 04 04:39:35 crc kubenswrapper[4575]: I1004 04:39:35.728141 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3"} err="failed to get container status \"54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3\": rpc error: code = NotFound desc = could not find container \"54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3\": container with ID starting with 54dbf49bae28e6ff77e52aa20ab7682a7e612b11b43435a6a3c1a8a5f18ffae3 not found: ID does not exist" Oct 04 04:39:36 crc kubenswrapper[4575]: I1004 04:39:36.771775 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zljdn"] Oct 04 04:39:36 crc kubenswrapper[4575]: I1004 04:39:36.772340 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zljdn" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerName="registry-server" containerID="cri-o://af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216" gracePeriod=2 Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.317697 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7816959d-a0f3-4093-aafa-b16835682a09" path="/var/lib/kubelet/pods/7816959d-a0f3-4093-aafa-b16835682a09/volumes" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.594298 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.673239 4575 generic.go:334] "Generic (PLEG): container finished" podID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerID="af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216" exitCode=0 Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.673304 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zljdn" event={"ID":"a5ffc246-1302-487b-bcaf-4aee56ed6534","Type":"ContainerDied","Data":"af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216"} Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.673336 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zljdn" event={"ID":"a5ffc246-1302-487b-bcaf-4aee56ed6534","Type":"ContainerDied","Data":"6bb4f7d8b0e8f256d84ed3e66872e93b5a6ea884bbdba6acf8c4c2104d5a5cfc"} Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.673358 4575 scope.go:117] "RemoveContainer" containerID="af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.673460 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zljdn" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.688140 4575 scope.go:117] "RemoveContainer" containerID="72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.698328 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-catalog-content\") pod \"a5ffc246-1302-487b-bcaf-4aee56ed6534\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.698372 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-utilities\") pod \"a5ffc246-1302-487b-bcaf-4aee56ed6534\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.698452 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx5l7\" (UniqueName: \"kubernetes.io/projected/a5ffc246-1302-487b-bcaf-4aee56ed6534-kube-api-access-rx5l7\") pod \"a5ffc246-1302-487b-bcaf-4aee56ed6534\" (UID: \"a5ffc246-1302-487b-bcaf-4aee56ed6534\") " Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.699065 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-utilities" (OuterVolumeSpecName: "utilities") pod "a5ffc246-1302-487b-bcaf-4aee56ed6534" (UID: "a5ffc246-1302-487b-bcaf-4aee56ed6534"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.705133 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5ffc246-1302-487b-bcaf-4aee56ed6534-kube-api-access-rx5l7" (OuterVolumeSpecName: "kube-api-access-rx5l7") pod "a5ffc246-1302-487b-bcaf-4aee56ed6534" (UID: "a5ffc246-1302-487b-bcaf-4aee56ed6534"). InnerVolumeSpecName "kube-api-access-rx5l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.706151 4575 scope.go:117] "RemoveContainer" containerID="687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.734074 4575 scope.go:117] "RemoveContainer" containerID="af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216" Oct 04 04:39:37 crc kubenswrapper[4575]: E1004 04:39:37.734487 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216\": container with ID starting with af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216 not found: ID does not exist" containerID="af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.734524 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216"} err="failed to get container status \"af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216\": rpc error: code = NotFound desc = could not find container \"af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216\": container with ID starting with af7f7bdd64f76570e926da1e4e95ec93602619726ee40cfc4d3475859a791216 not found: ID does not exist" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.734550 4575 scope.go:117] "RemoveContainer" containerID="72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78" Oct 04 04:39:37 crc kubenswrapper[4575]: E1004 04:39:37.735022 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78\": container with ID starting with 72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78 not found: ID does not exist" containerID="72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.735047 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78"} err="failed to get container status \"72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78\": rpc error: code = NotFound desc = could not find container \"72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78\": container with ID starting with 72c2f1958526997ce6585011c7f762ae59375c930adf92894175dff5fa0fed78 not found: ID does not exist" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.735062 4575 scope.go:117] "RemoveContainer" containerID="687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e" Oct 04 04:39:37 crc kubenswrapper[4575]: E1004 04:39:37.735388 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e\": container with ID starting with 687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e not found: ID does not exist" containerID="687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.735438 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e"} err="failed to get container status \"687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e\": rpc error: code = NotFound desc = could not find container \"687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e\": container with ID starting with 687b2640a0ea7f60fc9ba84df26ea3b4b8373c476e91bed39c5318f43b00761e not found: ID does not exist" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.782698 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5ffc246-1302-487b-bcaf-4aee56ed6534" (UID: "a5ffc246-1302-487b-bcaf-4aee56ed6534"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.799714 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx5l7\" (UniqueName: \"kubernetes.io/projected/a5ffc246-1302-487b-bcaf-4aee56ed6534-kube-api-access-rx5l7\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.799764 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:37 crc kubenswrapper[4575]: I1004 04:39:37.799781 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffc246-1302-487b-bcaf-4aee56ed6534-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:38 crc kubenswrapper[4575]: I1004 04:39:38.000241 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zljdn"] Oct 04 04:39:38 crc kubenswrapper[4575]: I1004 04:39:38.001985 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zljdn"] Oct 04 04:39:39 crc kubenswrapper[4575]: I1004 04:39:39.316964 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" path="/var/lib/kubelet/pods/a5ffc246-1302-487b-bcaf-4aee56ed6534/volumes" Oct 04 04:39:40 crc kubenswrapper[4575]: I1004 04:39:40.485451 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wk2s"] Oct 04 04:39:54 crc kubenswrapper[4575]: I1004 04:39:54.100364 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bmzcj" Oct 04 04:39:54 crc kubenswrapper[4575]: I1004 04:39:54.148331 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h5dwl"] Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.518772 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" podUID="952902f2-2371-4a82-84e7-365978f206f6" containerName="oauth-openshift" containerID="cri-o://7a921e49a9dafc55a414db61be8942c3d244f9c0754ff932610c9a895d17b23a" gracePeriod=15 Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.824967 4575 generic.go:334] "Generic (PLEG): container finished" podID="952902f2-2371-4a82-84e7-365978f206f6" containerID="7a921e49a9dafc55a414db61be8942c3d244f9c0754ff932610c9a895d17b23a" exitCode=0 Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.825060 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" event={"ID":"952902f2-2371-4a82-84e7-365978f206f6","Type":"ContainerDied","Data":"7a921e49a9dafc55a414db61be8942c3d244f9c0754ff932610c9a895d17b23a"} Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.825306 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" event={"ID":"952902f2-2371-4a82-84e7-365978f206f6","Type":"ContainerDied","Data":"32e1318cf7cc1f85162a7838853aef966e4a032b09e33a1efa7a91fd69944bc9"} Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.825325 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32e1318cf7cc1f85162a7838853aef966e4a032b09e33a1efa7a91fd69944bc9" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.839079 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.904647 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs"] Oct 04 04:40:05 crc kubenswrapper[4575]: E1004 04:40:05.904946 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952902f2-2371-4a82-84e7-365978f206f6" containerName="oauth-openshift" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.904968 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="952902f2-2371-4a82-84e7-365978f206f6" containerName="oauth-openshift" Oct 04 04:40:05 crc kubenswrapper[4575]: E1004 04:40:05.904983 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7816959d-a0f3-4093-aafa-b16835682a09" containerName="extract-content" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.904991 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7816959d-a0f3-4093-aafa-b16835682a09" containerName="extract-content" Oct 04 04:40:05 crc kubenswrapper[4575]: E1004 04:40:05.905002 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerName="registry-server" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.905010 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerName="registry-server" Oct 04 04:40:05 crc kubenswrapper[4575]: E1004 04:40:05.905023 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerName="extract-utilities" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.905032 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerName="extract-utilities" Oct 04 04:40:05 crc kubenswrapper[4575]: E1004 04:40:05.905046 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7816959d-a0f3-4093-aafa-b16835682a09" containerName="registry-server" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.905054 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7816959d-a0f3-4093-aafa-b16835682a09" containerName="registry-server" Oct 04 04:40:05 crc kubenswrapper[4575]: E1004 04:40:05.905065 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerName="extract-content" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.905073 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerName="extract-content" Oct 04 04:40:05 crc kubenswrapper[4575]: E1004 04:40:05.905089 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7816959d-a0f3-4093-aafa-b16835682a09" containerName="extract-utilities" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.905098 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7816959d-a0f3-4093-aafa-b16835682a09" containerName="extract-utilities" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.905211 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="7816959d-a0f3-4093-aafa-b16835682a09" containerName="registry-server" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.905228 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5ffc246-1302-487b-bcaf-4aee56ed6534" containerName="registry-server" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.905242 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="952902f2-2371-4a82-84e7-365978f206f6" containerName="oauth-openshift" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.905778 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.907818 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs"] Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.942679 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-ocp-branding-template\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.942884 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g45h5\" (UniqueName: \"kubernetes.io/projected/952902f2-2371-4a82-84e7-365978f206f6-kube-api-access-g45h5\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943229 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-trusted-ca-bundle\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943265 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-session\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943287 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-serving-cert\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943315 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-provider-selection\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943333 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/952902f2-2371-4a82-84e7-365978f206f6-audit-dir\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943358 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-audit-policies\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943379 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-login\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943404 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-idp-0-file-data\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943420 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-error\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943439 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-service-ca\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943457 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-router-certs\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943473 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-cliconfig\") pod \"952902f2-2371-4a82-84e7-365978f206f6\" (UID: \"952902f2-2371-4a82-84e7-365978f206f6\") " Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943575 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943629 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943649 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-template-login\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943667 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943693 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943717 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfhft\" (UniqueName: \"kubernetes.io/projected/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-kube-api-access-tfhft\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943743 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-service-ca\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943768 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-router-certs\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943795 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-template-error\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943821 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943868 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-audit-dir\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943892 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-audit-policies\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943912 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.943938 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-session\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.944505 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.944962 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/952902f2-2371-4a82-84e7-365978f206f6-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.945245 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.945488 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.945722 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.949042 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.949408 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.950849 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.950908 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952902f2-2371-4a82-84e7-365978f206f6-kube-api-access-g45h5" (OuterVolumeSpecName: "kube-api-access-g45h5") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "kube-api-access-g45h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.951276 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.951495 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.951646 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.951941 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:05 crc kubenswrapper[4575]: I1004 04:40:05.952122 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "952902f2-2371-4a82-84e7-365978f206f6" (UID: "952902f2-2371-4a82-84e7-365978f206f6"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044637 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044679 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044701 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-template-login\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044720 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044740 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044756 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfhft\" (UniqueName: \"kubernetes.io/projected/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-kube-api-access-tfhft\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044772 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-service-ca\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044788 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-router-certs\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044810 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-template-error\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044829 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044865 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-audit-dir\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044883 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-audit-policies\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044899 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044919 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-session\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044962 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044973 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044982 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.044992 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045002 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045011 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045021 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045029 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g45h5\" (UniqueName: \"kubernetes.io/projected/952902f2-2371-4a82-84e7-365978f206f6-kube-api-access-g45h5\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045038 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045048 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045058 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045073 4575 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/952902f2-2371-4a82-84e7-365978f206f6-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045086 4575 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/952902f2-2371-4a82-84e7-365978f206f6-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045097 4575 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/952902f2-2371-4a82-84e7-365978f206f6-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.045870 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-service-ca\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.046621 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.046685 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.047305 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-audit-policies\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.047364 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-audit-dir\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.047690 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-session\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.047817 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.047916 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.048879 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.050036 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.050679 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-system-router-certs\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.051140 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-template-error\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.051185 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-v4-0-config-user-template-login\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.062348 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfhft\" (UniqueName: \"kubernetes.io/projected/08bf79c0-187a-480c-a2be-3c1a99a5ab4d-kube-api-access-tfhft\") pod \"oauth-openshift-74fb9cdb46-pzlhs\" (UID: \"08bf79c0-187a-480c-a2be-3c1a99a5ab4d\") " pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.221830 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.399849 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs"] Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.832903 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2wk2s" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.834767 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" event={"ID":"08bf79c0-187a-480c-a2be-3c1a99a5ab4d","Type":"ContainerStarted","Data":"a86f1d8a9182a3cb437d95170658a9942ab92258babc05a9e2d1e7225f6d7b69"} Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.834809 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" event={"ID":"08bf79c0-187a-480c-a2be-3c1a99a5ab4d","Type":"ContainerStarted","Data":"bd90960345095b20e9ce2eec35972cd2a57bd872b38c5b8372ba782e4aeb291e"} Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.835177 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.860996 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" podStartSLOduration=26.860976889 podStartE2EDuration="26.860976889s" podCreationTimestamp="2025-10-04 04:39:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:40:06.859791804 +0000 UTC m=+358.188350618" watchObservedRunningTime="2025-10-04 04:40:06.860976889 +0000 UTC m=+358.189535703" Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.871392 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wk2s"] Oct 04 04:40:06 crc kubenswrapper[4575]: I1004 04:40:06.877894 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2wk2s"] Oct 04 04:40:07 crc kubenswrapper[4575]: I1004 04:40:07.064482 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-74fb9cdb46-pzlhs" Oct 04 04:40:07 crc kubenswrapper[4575]: I1004 04:40:07.319372 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="952902f2-2371-4a82-84e7-365978f206f6" path="/var/lib/kubelet/pods/952902f2-2371-4a82-84e7-365978f206f6/volumes" Oct 04 04:40:14 crc kubenswrapper[4575]: I1004 04:40:14.967169 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mkz66"] Oct 04 04:40:14 crc kubenswrapper[4575]: I1004 04:40:14.968204 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mkz66" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerName="registry-server" containerID="cri-o://4c6a7d047e294e4598b55c4686e14eef2aa50b089cb2a3e926c6b8f21eb738a7" gracePeriod=30 Oct 04 04:40:14 crc kubenswrapper[4575]: I1004 04:40:14.970927 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x4zr9"] Oct 04 04:40:14 crc kubenswrapper[4575]: I1004 04:40:14.971212 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x4zr9" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerName="registry-server" containerID="cri-o://f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3" gracePeriod=30 Oct 04 04:40:14 crc kubenswrapper[4575]: I1004 04:40:14.976262 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fn6w5"] Oct 04 04:40:14 crc kubenswrapper[4575]: I1004 04:40:14.976861 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" podUID="844cd2d0-93d0-45d9-b342-b86df49548d8" containerName="marketplace-operator" containerID="cri-o://280ba2abaf1b19ae3d72e9df4ce5ac0e9096e9de2239b8ee86023490b914e23b" gracePeriod=30 Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.004246 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdbvl"] Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.004567 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rdbvl" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" containerName="registry-server" containerID="cri-o://a086d332af4c29416e9fefa267029dea60413a85427e03abb322dc592e23980f" gracePeriod=30 Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.015340 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66vsx"] Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.015639 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-66vsx" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerName="registry-server" containerID="cri-o://16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab" gracePeriod=30 Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.020606 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7nh7v"] Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.023881 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.050401 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7nh7v"] Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.068877 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85fqg\" (UniqueName: \"kubernetes.io/projected/1f7d3832-1561-4d4f-a47a-311b5aadef90-kube-api-access-85fqg\") pod \"marketplace-operator-79b997595-7nh7v\" (UID: \"1f7d3832-1561-4d4f-a47a-311b5aadef90\") " pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.069129 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f7d3832-1561-4d4f-a47a-311b5aadef90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7nh7v\" (UID: \"1f7d3832-1561-4d4f-a47a-311b5aadef90\") " pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.069724 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1f7d3832-1561-4d4f-a47a-311b5aadef90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7nh7v\" (UID: \"1f7d3832-1561-4d4f-a47a-311b5aadef90\") " pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.171621 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f7d3832-1561-4d4f-a47a-311b5aadef90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7nh7v\" (UID: \"1f7d3832-1561-4d4f-a47a-311b5aadef90\") " pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.172156 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1f7d3832-1561-4d4f-a47a-311b5aadef90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7nh7v\" (UID: \"1f7d3832-1561-4d4f-a47a-311b5aadef90\") " pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.172308 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85fqg\" (UniqueName: \"kubernetes.io/projected/1f7d3832-1561-4d4f-a47a-311b5aadef90-kube-api-access-85fqg\") pod \"marketplace-operator-79b997595-7nh7v\" (UID: \"1f7d3832-1561-4d4f-a47a-311b5aadef90\") " pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.173180 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f7d3832-1561-4d4f-a47a-311b5aadef90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7nh7v\" (UID: \"1f7d3832-1561-4d4f-a47a-311b5aadef90\") " pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.181313 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1f7d3832-1561-4d4f-a47a-311b5aadef90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7nh7v\" (UID: \"1f7d3832-1561-4d4f-a47a-311b5aadef90\") " pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.195710 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85fqg\" (UniqueName: \"kubernetes.io/projected/1f7d3832-1561-4d4f-a47a-311b5aadef90-kube-api-access-85fqg\") pod \"marketplace-operator-79b997595-7nh7v\" (UID: \"1f7d3832-1561-4d4f-a47a-311b5aadef90\") " pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.342930 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.485749 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.553224 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7nh7v"] Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.676949 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgcpf\" (UniqueName: \"kubernetes.io/projected/80c6efe3-33be-46c4-b557-ea8fbfed349a-kube-api-access-mgcpf\") pod \"80c6efe3-33be-46c4-b557-ea8fbfed349a\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.677289 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-utilities\") pod \"80c6efe3-33be-46c4-b557-ea8fbfed349a\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.677328 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-catalog-content\") pod \"80c6efe3-33be-46c4-b557-ea8fbfed349a\" (UID: \"80c6efe3-33be-46c4-b557-ea8fbfed349a\") " Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.678459 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-utilities" (OuterVolumeSpecName: "utilities") pod "80c6efe3-33be-46c4-b557-ea8fbfed349a" (UID: "80c6efe3-33be-46c4-b557-ea8fbfed349a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.682200 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80c6efe3-33be-46c4-b557-ea8fbfed349a-kube-api-access-mgcpf" (OuterVolumeSpecName: "kube-api-access-mgcpf") pod "80c6efe3-33be-46c4-b557-ea8fbfed349a" (UID: "80c6efe3-33be-46c4-b557-ea8fbfed349a"). InnerVolumeSpecName "kube-api-access-mgcpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.762046 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80c6efe3-33be-46c4-b557-ea8fbfed349a" (UID: "80c6efe3-33be-46c4-b557-ea8fbfed349a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.778561 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgcpf\" (UniqueName: \"kubernetes.io/projected/80c6efe3-33be-46c4-b557-ea8fbfed349a-kube-api-access-mgcpf\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.778617 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.778632 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c6efe3-33be-46c4-b557-ea8fbfed349a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.855936 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.878906 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84dd2\" (UniqueName: \"kubernetes.io/projected/f9b027e7-fd36-43e5-80b2-53e105822d71-kube-api-access-84dd2\") pod \"f9b027e7-fd36-43e5-80b2-53e105822d71\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.879001 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-utilities\") pod \"f9b027e7-fd36-43e5-80b2-53e105822d71\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.879023 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-catalog-content\") pod \"f9b027e7-fd36-43e5-80b2-53e105822d71\" (UID: \"f9b027e7-fd36-43e5-80b2-53e105822d71\") " Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.882467 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-utilities" (OuterVolumeSpecName: "utilities") pod "f9b027e7-fd36-43e5-80b2-53e105822d71" (UID: "f9b027e7-fd36-43e5-80b2-53e105822d71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.885170 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9b027e7-fd36-43e5-80b2-53e105822d71-kube-api-access-84dd2" (OuterVolumeSpecName: "kube-api-access-84dd2") pod "f9b027e7-fd36-43e5-80b2-53e105822d71" (UID: "f9b027e7-fd36-43e5-80b2-53e105822d71"). InnerVolumeSpecName "kube-api-access-84dd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.897489 4575 generic.go:334] "Generic (PLEG): container finished" podID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerID="f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3" exitCode=0 Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.897556 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4zr9" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.897618 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4zr9" event={"ID":"f9b027e7-fd36-43e5-80b2-53e105822d71","Type":"ContainerDied","Data":"f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3"} Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.897654 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4zr9" event={"ID":"f9b027e7-fd36-43e5-80b2-53e105822d71","Type":"ContainerDied","Data":"2b0813fdb72839fc558f5fb48611b83d38defcce8ce02b8fb8c72e9ca9bd8e34"} Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.897673 4575 scope.go:117] "RemoveContainer" containerID="f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.904332 4575 generic.go:334] "Generic (PLEG): container finished" podID="9a908c20-e144-41e6-b331-f5972caf9a33" containerID="a086d332af4c29416e9fefa267029dea60413a85427e03abb322dc592e23980f" exitCode=0 Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.904400 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdbvl" event={"ID":"9a908c20-e144-41e6-b331-f5972caf9a33","Type":"ContainerDied","Data":"a086d332af4c29416e9fefa267029dea60413a85427e03abb322dc592e23980f"} Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.916469 4575 generic.go:334] "Generic (PLEG): container finished" podID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerID="16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab" exitCode=0 Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.916578 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vsx" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.916561 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vsx" event={"ID":"80c6efe3-33be-46c4-b557-ea8fbfed349a","Type":"ContainerDied","Data":"16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab"} Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.916661 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vsx" event={"ID":"80c6efe3-33be-46c4-b557-ea8fbfed349a","Type":"ContainerDied","Data":"e8bfc2936f4ff8db8ec8e402872ce76e82ce4fd6d1c4a905053dc61ffa5ac7f4"} Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.923671 4575 scope.go:117] "RemoveContainer" containerID="4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.924274 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9b027e7-fd36-43e5-80b2-53e105822d71" (UID: "f9b027e7-fd36-43e5-80b2-53e105822d71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.925042 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" event={"ID":"1f7d3832-1561-4d4f-a47a-311b5aadef90","Type":"ContainerStarted","Data":"9bc14f7feebaa91a640c4aefd0ffddaeec1bc70a39b4b9f5b57beccdf512aceb"} Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.925473 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" event={"ID":"1f7d3832-1561-4d4f-a47a-311b5aadef90","Type":"ContainerStarted","Data":"5e61a0daee7a21ee42fafc04081234a2edf38015f7896801d1f22afcfdc44c7f"} Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.930746 4575 generic.go:334] "Generic (PLEG): container finished" podID="844cd2d0-93d0-45d9-b342-b86df49548d8" containerID="280ba2abaf1b19ae3d72e9df4ce5ac0e9096e9de2239b8ee86023490b914e23b" exitCode=0 Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.930969 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" event={"ID":"844cd2d0-93d0-45d9-b342-b86df49548d8","Type":"ContainerDied","Data":"280ba2abaf1b19ae3d72e9df4ce5ac0e9096e9de2239b8ee86023490b914e23b"} Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.934260 4575 generic.go:334] "Generic (PLEG): container finished" podID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerID="4c6a7d047e294e4598b55c4686e14eef2aa50b089cb2a3e926c6b8f21eb738a7" exitCode=0 Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.934323 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkz66" event={"ID":"7846a3af-bfa1-4129-ae19-48bc580c0a4a","Type":"ContainerDied","Data":"4c6a7d047e294e4598b55c4686e14eef2aa50b089cb2a3e926c6b8f21eb738a7"} Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.979999 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84dd2\" (UniqueName: \"kubernetes.io/projected/f9b027e7-fd36-43e5-80b2-53e105822d71-kube-api-access-84dd2\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.980034 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:15 crc kubenswrapper[4575]: I1004 04:40:15.980043 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b027e7-fd36-43e5-80b2-53e105822d71-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.011230 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66vsx"] Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.011284 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-66vsx"] Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.015160 4575 scope.go:117] "RemoveContainer" containerID="46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.090416 4575 scope.go:117] "RemoveContainer" containerID="f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3" Oct 04 04:40:16 crc kubenswrapper[4575]: E1004 04:40:16.090970 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3\": container with ID starting with f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3 not found: ID does not exist" containerID="f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.091002 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3"} err="failed to get container status \"f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3\": rpc error: code = NotFound desc = could not find container \"f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3\": container with ID starting with f771c497592e5753400317e7ac4cec9e425bf17ac34779efe8d111c0cd7753d3 not found: ID does not exist" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.091028 4575 scope.go:117] "RemoveContainer" containerID="4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26" Oct 04 04:40:16 crc kubenswrapper[4575]: E1004 04:40:16.091446 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26\": container with ID starting with 4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26 not found: ID does not exist" containerID="4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.091473 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26"} err="failed to get container status \"4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26\": rpc error: code = NotFound desc = could not find container \"4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26\": container with ID starting with 4650326b338bc5949e74a60fd52dee205efe5af53185ab39aa04850a9fb3ed26 not found: ID does not exist" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.091489 4575 scope.go:117] "RemoveContainer" containerID="46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b" Oct 04 04:40:16 crc kubenswrapper[4575]: E1004 04:40:16.091830 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b\": container with ID starting with 46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b not found: ID does not exist" containerID="46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.091853 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b"} err="failed to get container status \"46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b\": rpc error: code = NotFound desc = could not find container \"46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b\": container with ID starting with 46f5d452084cc452eb309658b3421cffa0bff4f63fe8d930fd260c4299f4606b not found: ID does not exist" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.091866 4575 scope.go:117] "RemoveContainer" containerID="16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.169576 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.174388 4575 scope.go:117] "RemoveContainer" containerID="2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.174785 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.184138 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-utilities\") pod \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.184200 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-trusted-ca\") pod \"844cd2d0-93d0-45d9-b342-b86df49548d8\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.184247 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfz6h\" (UniqueName: \"kubernetes.io/projected/7846a3af-bfa1-4129-ae19-48bc580c0a4a-kube-api-access-nfz6h\") pod \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.184895 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.185350 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "844cd2d0-93d0-45d9-b342-b86df49548d8" (UID: "844cd2d0-93d0-45d9-b342-b86df49548d8"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.187522 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7846a3af-bfa1-4129-ae19-48bc580c0a4a-kube-api-access-nfz6h" (OuterVolumeSpecName: "kube-api-access-nfz6h") pod "7846a3af-bfa1-4129-ae19-48bc580c0a4a" (UID: "7846a3af-bfa1-4129-ae19-48bc580c0a4a"). InnerVolumeSpecName "kube-api-access-nfz6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.189608 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-utilities" (OuterVolumeSpecName: "utilities") pod "7846a3af-bfa1-4129-ae19-48bc580c0a4a" (UID: "7846a3af-bfa1-4129-ae19-48bc580c0a4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.200971 4575 scope.go:117] "RemoveContainer" containerID="e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.272633 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x4zr9"] Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.273086 4575 scope.go:117] "RemoveContainer" containerID="16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab" Oct 04 04:40:16 crc kubenswrapper[4575]: E1004 04:40:16.274732 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab\": container with ID starting with 16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab not found: ID does not exist" containerID="16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.274776 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab"} err="failed to get container status \"16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab\": rpc error: code = NotFound desc = could not find container \"16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab\": container with ID starting with 16d02492d2f2966cd907f5aa3ad705ccb308b8ad765fe58982020acdf646adab not found: ID does not exist" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.274810 4575 scope.go:117] "RemoveContainer" containerID="2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49" Oct 04 04:40:16 crc kubenswrapper[4575]: E1004 04:40:16.275106 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49\": container with ID starting with 2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49 not found: ID does not exist" containerID="2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.275142 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49"} err="failed to get container status \"2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49\": rpc error: code = NotFound desc = could not find container \"2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49\": container with ID starting with 2d015105629764ddf6a33c73bc425ab117a19355619e9d0651e0c8ce64b52d49 not found: ID does not exist" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.275174 4575 scope.go:117] "RemoveContainer" containerID="e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd" Oct 04 04:40:16 crc kubenswrapper[4575]: E1004 04:40:16.275919 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd\": container with ID starting with e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd not found: ID does not exist" containerID="e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.275946 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd"} err="failed to get container status \"e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd\": rpc error: code = NotFound desc = could not find container \"e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd\": container with ID starting with e36e4fffc114f139eeb40f09207698f9cd531a850f9af35434238acfacc7ddbd not found: ID does not exist" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.285085 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x4zr9"] Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.285437 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-catalog-content\") pod \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\" (UID: \"7846a3af-bfa1-4129-ae19-48bc580c0a4a\") " Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.285580 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-operator-metrics\") pod \"844cd2d0-93d0-45d9-b342-b86df49548d8\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.285741 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9zv2\" (UniqueName: \"kubernetes.io/projected/844cd2d0-93d0-45d9-b342-b86df49548d8-kube-api-access-m9zv2\") pod \"844cd2d0-93d0-45d9-b342-b86df49548d8\" (UID: \"844cd2d0-93d0-45d9-b342-b86df49548d8\") " Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.286167 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.286210 4575 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.286224 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfz6h\" (UniqueName: \"kubernetes.io/projected/7846a3af-bfa1-4129-ae19-48bc580c0a4a-kube-api-access-nfz6h\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.289300 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/844cd2d0-93d0-45d9-b342-b86df49548d8-kube-api-access-m9zv2" (OuterVolumeSpecName: "kube-api-access-m9zv2") pod "844cd2d0-93d0-45d9-b342-b86df49548d8" (UID: "844cd2d0-93d0-45d9-b342-b86df49548d8"). InnerVolumeSpecName "kube-api-access-m9zv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.294956 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "844cd2d0-93d0-45d9-b342-b86df49548d8" (UID: "844cd2d0-93d0-45d9-b342-b86df49548d8"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.339459 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7846a3af-bfa1-4129-ae19-48bc580c0a4a" (UID: "7846a3af-bfa1-4129-ae19-48bc580c0a4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.387036 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-catalog-content\") pod \"9a908c20-e144-41e6-b331-f5972caf9a33\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.387143 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-utilities\") pod \"9a908c20-e144-41e6-b331-f5972caf9a33\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.387186 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kdwq\" (UniqueName: \"kubernetes.io/projected/9a908c20-e144-41e6-b331-f5972caf9a33-kube-api-access-7kdwq\") pod \"9a908c20-e144-41e6-b331-f5972caf9a33\" (UID: \"9a908c20-e144-41e6-b331-f5972caf9a33\") " Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.387507 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9zv2\" (UniqueName: \"kubernetes.io/projected/844cd2d0-93d0-45d9-b342-b86df49548d8-kube-api-access-m9zv2\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.387528 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7846a3af-bfa1-4129-ae19-48bc580c0a4a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.387543 4575 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/844cd2d0-93d0-45d9-b342-b86df49548d8-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.388565 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-utilities" (OuterVolumeSpecName: "utilities") pod "9a908c20-e144-41e6-b331-f5972caf9a33" (UID: "9a908c20-e144-41e6-b331-f5972caf9a33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.390318 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a908c20-e144-41e6-b331-f5972caf9a33-kube-api-access-7kdwq" (OuterVolumeSpecName: "kube-api-access-7kdwq") pod "9a908c20-e144-41e6-b331-f5972caf9a33" (UID: "9a908c20-e144-41e6-b331-f5972caf9a33"). InnerVolumeSpecName "kube-api-access-7kdwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.399838 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a908c20-e144-41e6-b331-f5972caf9a33" (UID: "9a908c20-e144-41e6-b331-f5972caf9a33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.489211 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.489255 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a908c20-e144-41e6-b331-f5972caf9a33-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.489267 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kdwq\" (UniqueName: \"kubernetes.io/projected/9a908c20-e144-41e6-b331-f5972caf9a33-kube-api-access-7kdwq\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.942571 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdbvl" event={"ID":"9a908c20-e144-41e6-b331-f5972caf9a33","Type":"ContainerDied","Data":"0c9aa185fbde3548114c490edce6591c1504e227893ec467a04283a314f3b3ff"} Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.942876 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdbvl" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.942916 4575 scope.go:117] "RemoveContainer" containerID="a086d332af4c29416e9fefa267029dea60413a85427e03abb322dc592e23980f" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.947728 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" event={"ID":"844cd2d0-93d0-45d9-b342-b86df49548d8","Type":"ContainerDied","Data":"5e469dfa42bf7816bd45841f333dd23174ae7e94cac2591f2a9e7032018096fb"} Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.947845 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fn6w5" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.953607 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkz66" event={"ID":"7846a3af-bfa1-4129-ae19-48bc580c0a4a","Type":"ContainerDied","Data":"11a8fd1d569d3442d78d53d1519808fcbff7d8139793924abbd6ac3882101eaa"} Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.953968 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkz66" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.956188 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.961103 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.962209 4575 scope.go:117] "RemoveContainer" containerID="8a7edd9d85fbde7cdefd4cad1fafb6f0b9b21c71d7780f6c8f88cdf91a780aab" Oct 04 04:40:16 crc kubenswrapper[4575]: I1004 04:40:16.989717 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7nh7v" podStartSLOduration=2.989690837 podStartE2EDuration="2.989690837s" podCreationTimestamp="2025-10-04 04:40:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:40:16.983124462 +0000 UTC m=+368.311683296" watchObservedRunningTime="2025-10-04 04:40:16.989690837 +0000 UTC m=+368.318249661" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.008116 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fn6w5"] Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.011311 4575 scope.go:117] "RemoveContainer" containerID="e10154aa13a9517746a67a2d406e250704f4a38102c0ecc7d1a23eea9fccc90b" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.013055 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fn6w5"] Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.027967 4575 scope.go:117] "RemoveContainer" containerID="280ba2abaf1b19ae3d72e9df4ce5ac0e9096e9de2239b8ee86023490b914e23b" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.031768 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mkz66"] Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.034040 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mkz66"] Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.049744 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdbvl"] Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.055449 4575 scope.go:117] "RemoveContainer" containerID="4c6a7d047e294e4598b55c4686e14eef2aa50b089cb2a3e926c6b8f21eb738a7" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.060202 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdbvl"] Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.080372 4575 scope.go:117] "RemoveContainer" containerID="a974e424ae7b55d181273802ef21b632423d4a0f42e9db87601e63dc42330403" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.122072 4575 scope.go:117] "RemoveContainer" containerID="59a575ec21f05713d562e47485822d48fa16b53cae915693a0a06a0b93abcd97" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.315167 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" path="/var/lib/kubelet/pods/7846a3af-bfa1-4129-ae19-48bc580c0a4a/volumes" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.315834 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" path="/var/lib/kubelet/pods/80c6efe3-33be-46c4-b557-ea8fbfed349a/volumes" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.316662 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="844cd2d0-93d0-45d9-b342-b86df49548d8" path="/var/lib/kubelet/pods/844cd2d0-93d0-45d9-b342-b86df49548d8/volumes" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.317790 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" path="/var/lib/kubelet/pods/9a908c20-e144-41e6-b331-f5972caf9a33/volumes" Oct 04 04:40:17 crc kubenswrapper[4575]: I1004 04:40:17.318436 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" path="/var/lib/kubelet/pods/f9b027e7-fd36-43e5-80b2-53e105822d71/volumes" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.175559 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qfp6s"] Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176073 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" containerName="extract-content" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176087 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" containerName="extract-content" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176104 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerName="extract-utilities" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176112 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerName="extract-utilities" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176120 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176127 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176135 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176140 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176148 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerName="extract-content" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176154 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerName="extract-content" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176164 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerName="extract-utilities" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176181 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerName="extract-utilities" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176189 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176196 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176208 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="844cd2d0-93d0-45d9-b342-b86df49548d8" containerName="marketplace-operator" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176216 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="844cd2d0-93d0-45d9-b342-b86df49548d8" containerName="marketplace-operator" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176227 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" containerName="extract-utilities" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176234 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" containerName="extract-utilities" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176245 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerName="extract-utilities" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176254 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerName="extract-utilities" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176265 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176272 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176282 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerName="extract-content" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176290 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerName="extract-content" Oct 04 04:40:19 crc kubenswrapper[4575]: E1004 04:40:19.176299 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerName="extract-content" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176305 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerName="extract-content" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176408 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9b027e7-fd36-43e5-80b2-53e105822d71" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176425 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="80c6efe3-33be-46c4-b557-ea8fbfed349a" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176435 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="7846a3af-bfa1-4129-ae19-48bc580c0a4a" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176448 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="844cd2d0-93d0-45d9-b342-b86df49548d8" containerName="marketplace-operator" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.176457 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a908c20-e144-41e6-b331-f5972caf9a33" containerName="registry-server" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.177350 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.179719 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.188655 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qfp6s"] Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.191696 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" podUID="2b7b2fb7-8708-45e5-8603-3e7c13f020f8" containerName="registry" containerID="cri-o://7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7" gracePeriod=30 Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.340682 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll9rm\" (UniqueName: \"kubernetes.io/projected/cbaf719e-0879-45d7-a285-42109d7fb978-kube-api-access-ll9rm\") pod \"redhat-operators-qfp6s\" (UID: \"cbaf719e-0879-45d7-a285-42109d7fb978\") " pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.340741 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbaf719e-0879-45d7-a285-42109d7fb978-utilities\") pod \"redhat-operators-qfp6s\" (UID: \"cbaf719e-0879-45d7-a285-42109d7fb978\") " pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.340792 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbaf719e-0879-45d7-a285-42109d7fb978-catalog-content\") pod \"redhat-operators-qfp6s\" (UID: \"cbaf719e-0879-45d7-a285-42109d7fb978\") " pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.384033 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ktd5v"] Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.393036 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.393043 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ktd5v"] Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.397941 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.441726 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbaf719e-0879-45d7-a285-42109d7fb978-catalog-content\") pod \"redhat-operators-qfp6s\" (UID: \"cbaf719e-0879-45d7-a285-42109d7fb978\") " pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.441956 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll9rm\" (UniqueName: \"kubernetes.io/projected/cbaf719e-0879-45d7-a285-42109d7fb978-kube-api-access-ll9rm\") pod \"redhat-operators-qfp6s\" (UID: \"cbaf719e-0879-45d7-a285-42109d7fb978\") " pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.442000 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbaf719e-0879-45d7-a285-42109d7fb978-utilities\") pod \"redhat-operators-qfp6s\" (UID: \"cbaf719e-0879-45d7-a285-42109d7fb978\") " pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.442319 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbaf719e-0879-45d7-a285-42109d7fb978-catalog-content\") pod \"redhat-operators-qfp6s\" (UID: \"cbaf719e-0879-45d7-a285-42109d7fb978\") " pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.442371 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbaf719e-0879-45d7-a285-42109d7fb978-utilities\") pod \"redhat-operators-qfp6s\" (UID: \"cbaf719e-0879-45d7-a285-42109d7fb978\") " pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.462082 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll9rm\" (UniqueName: \"kubernetes.io/projected/cbaf719e-0879-45d7-a285-42109d7fb978-kube-api-access-ll9rm\") pod \"redhat-operators-qfp6s\" (UID: \"cbaf719e-0879-45d7-a285-42109d7fb978\") " pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.499372 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.542795 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzr77\" (UniqueName: \"kubernetes.io/projected/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-kube-api-access-nzr77\") pod \"community-operators-ktd5v\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.543185 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-catalog-content\") pod \"community-operators-ktd5v\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.543223 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-utilities\") pod \"community-operators-ktd5v\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.606641 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.649278 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzr77\" (UniqueName: \"kubernetes.io/projected/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-kube-api-access-nzr77\") pod \"community-operators-ktd5v\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.649823 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-catalog-content\") pod \"community-operators-ktd5v\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.649877 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-utilities\") pod \"community-operators-ktd5v\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.650425 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-utilities\") pod \"community-operators-ktd5v\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.650427 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-catalog-content\") pod \"community-operators-ktd5v\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.675462 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzr77\" (UniqueName: \"kubernetes.io/projected/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-kube-api-access-nzr77\") pod \"community-operators-ktd5v\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.712885 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.737250 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qfp6s"] Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.750744 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-ca-trust-extracted\") pod \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.764789 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-trusted-ca\") pod \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.764864 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-installation-pull-secrets\") pod \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.765056 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-bound-sa-token\") pod \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.765094 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-tls\") pod \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.765193 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.765216 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-certificates\") pod \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.765236 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zqm8\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-kube-api-access-6zqm8\") pod \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\" (UID: \"2b7b2fb7-8708-45e5-8603-3e7c13f020f8\") " Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.765639 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2b7b2fb7-8708-45e5-8603-3e7c13f020f8" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:40:19 crc kubenswrapper[4575]: W1004 04:40:19.767144 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbaf719e_0879_45d7_a285_42109d7fb978.slice/crio-76f5fb6590a5d755738625e3328d75c561e74b852ade77aa0b34b4b1e8291532 WatchSource:0}: Error finding container 76f5fb6590a5d755738625e3328d75c561e74b852ade77aa0b34b4b1e8291532: Status 404 returned error can't find the container with id 76f5fb6590a5d755738625e3328d75c561e74b852ade77aa0b34b4b1e8291532 Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.767736 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2b7b2fb7-8708-45e5-8603-3e7c13f020f8" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.769286 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2b7b2fb7-8708-45e5-8603-3e7c13f020f8" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.769351 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2b7b2fb7-8708-45e5-8603-3e7c13f020f8" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.770313 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2b7b2fb7-8708-45e5-8603-3e7c13f020f8" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.770627 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2b7b2fb7-8708-45e5-8603-3e7c13f020f8" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.771047 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-kube-api-access-6zqm8" (OuterVolumeSpecName: "kube-api-access-6zqm8") pod "2b7b2fb7-8708-45e5-8603-3e7c13f020f8" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8"). InnerVolumeSpecName "kube-api-access-6zqm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.778075 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2b7b2fb7-8708-45e5-8603-3e7c13f020f8" (UID: "2b7b2fb7-8708-45e5-8603-3e7c13f020f8"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.866828 4575 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.867161 4575 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.867175 4575 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.867188 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zqm8\" (UniqueName: \"kubernetes.io/projected/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-kube-api-access-6zqm8\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.867199 4575 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.867211 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.867222 4575 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b7b2fb7-8708-45e5-8603-3e7c13f020f8-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.909539 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ktd5v"] Oct 04 04:40:19 crc kubenswrapper[4575]: W1004 04:40:19.919039 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6b72fdb_cfb9_4bb7_bb8e_2ffbfbbb8354.slice/crio-800ee2598d3d07d51d233d407952e1cb49ca30be90e05cc5bcf2f25d29b6f75a WatchSource:0}: Error finding container 800ee2598d3d07d51d233d407952e1cb49ca30be90e05cc5bcf2f25d29b6f75a: Status 404 returned error can't find the container with id 800ee2598d3d07d51d233d407952e1cb49ca30be90e05cc5bcf2f25d29b6f75a Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.977637 4575 generic.go:334] "Generic (PLEG): container finished" podID="cbaf719e-0879-45d7-a285-42109d7fb978" containerID="74771c608e76ec9c0437652e7a52a2a4a0076af4a4e30ba363ea9a54d20ba3b3" exitCode=0 Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.977742 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfp6s" event={"ID":"cbaf719e-0879-45d7-a285-42109d7fb978","Type":"ContainerDied","Data":"74771c608e76ec9c0437652e7a52a2a4a0076af4a4e30ba363ea9a54d20ba3b3"} Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.977798 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfp6s" event={"ID":"cbaf719e-0879-45d7-a285-42109d7fb978","Type":"ContainerStarted","Data":"76f5fb6590a5d755738625e3328d75c561e74b852ade77aa0b34b4b1e8291532"} Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.979227 4575 generic.go:334] "Generic (PLEG): container finished" podID="2b7b2fb7-8708-45e5-8603-3e7c13f020f8" containerID="7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7" exitCode=0 Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.979292 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" event={"ID":"2b7b2fb7-8708-45e5-8603-3e7c13f020f8","Type":"ContainerDied","Data":"7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7"} Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.979305 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.979323 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h5dwl" event={"ID":"2b7b2fb7-8708-45e5-8603-3e7c13f020f8","Type":"ContainerDied","Data":"60c85b8b74d19a9a0e5f865fc70b7a85331261c42953c850e5144cecd740aeeb"} Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.979344 4575 scope.go:117] "RemoveContainer" containerID="7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7" Oct 04 04:40:19 crc kubenswrapper[4575]: I1004 04:40:19.980999 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktd5v" event={"ID":"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354","Type":"ContainerStarted","Data":"800ee2598d3d07d51d233d407952e1cb49ca30be90e05cc5bcf2f25d29b6f75a"} Oct 04 04:40:20 crc kubenswrapper[4575]: I1004 04:40:20.022032 4575 scope.go:117] "RemoveContainer" containerID="7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7" Oct 04 04:40:20 crc kubenswrapper[4575]: E1004 04:40:20.022384 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7\": container with ID starting with 7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7 not found: ID does not exist" containerID="7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7" Oct 04 04:40:20 crc kubenswrapper[4575]: I1004 04:40:20.022413 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7"} err="failed to get container status \"7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7\": rpc error: code = NotFound desc = could not find container \"7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7\": container with ID starting with 7d4e153edef90c89c015910cd5dce06f41caa04314e8b5740849967c847429b7 not found: ID does not exist" Oct 04 04:40:20 crc kubenswrapper[4575]: I1004 04:40:20.044494 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h5dwl"] Oct 04 04:40:20 crc kubenswrapper[4575]: I1004 04:40:20.047772 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h5dwl"] Oct 04 04:40:20 crc kubenswrapper[4575]: I1004 04:40:20.987698 4575 generic.go:334] "Generic (PLEG): container finished" podID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerID="b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0" exitCode=0 Oct 04 04:40:20 crc kubenswrapper[4575]: I1004 04:40:20.987820 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktd5v" event={"ID":"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354","Type":"ContainerDied","Data":"b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0"} Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.315839 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b7b2fb7-8708-45e5-8603-3e7c13f020f8" path="/var/lib/kubelet/pods/2b7b2fb7-8708-45e5-8603-3e7c13f020f8/volumes" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.572793 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ncgzm"] Oct 04 04:40:21 crc kubenswrapper[4575]: E1004 04:40:21.572965 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7b2fb7-8708-45e5-8603-3e7c13f020f8" containerName="registry" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.572976 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7b2fb7-8708-45e5-8603-3e7c13f020f8" containerName="registry" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.573048 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b7b2fb7-8708-45e5-8603-3e7c13f020f8" containerName="registry" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.573647 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.577446 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.578353 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ncgzm"] Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.689606 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqx2z\" (UniqueName: \"kubernetes.io/projected/9f084f23-241a-45b2-88a9-2ffc29b161b4-kube-api-access-gqx2z\") pod \"redhat-marketplace-ncgzm\" (UID: \"9f084f23-241a-45b2-88a9-2ffc29b161b4\") " pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.689661 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f084f23-241a-45b2-88a9-2ffc29b161b4-utilities\") pod \"redhat-marketplace-ncgzm\" (UID: \"9f084f23-241a-45b2-88a9-2ffc29b161b4\") " pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.689773 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f084f23-241a-45b2-88a9-2ffc29b161b4-catalog-content\") pod \"redhat-marketplace-ncgzm\" (UID: \"9f084f23-241a-45b2-88a9-2ffc29b161b4\") " pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.770706 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9vz74"] Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.773147 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.775205 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.791521 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f084f23-241a-45b2-88a9-2ffc29b161b4-utilities\") pod \"redhat-marketplace-ncgzm\" (UID: \"9f084f23-241a-45b2-88a9-2ffc29b161b4\") " pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.791611 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aa15892-e136-4c58-aaf2-d62597704254-utilities\") pod \"certified-operators-9vz74\" (UID: \"1aa15892-e136-4c58-aaf2-d62597704254\") " pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.791662 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aa15892-e136-4c58-aaf2-d62597704254-catalog-content\") pod \"certified-operators-9vz74\" (UID: \"1aa15892-e136-4c58-aaf2-d62597704254\") " pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.791698 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f084f23-241a-45b2-88a9-2ffc29b161b4-catalog-content\") pod \"redhat-marketplace-ncgzm\" (UID: \"9f084f23-241a-45b2-88a9-2ffc29b161b4\") " pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.791728 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhrf4\" (UniqueName: \"kubernetes.io/projected/1aa15892-e136-4c58-aaf2-d62597704254-kube-api-access-nhrf4\") pod \"certified-operators-9vz74\" (UID: \"1aa15892-e136-4c58-aaf2-d62597704254\") " pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.791751 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqx2z\" (UniqueName: \"kubernetes.io/projected/9f084f23-241a-45b2-88a9-2ffc29b161b4-kube-api-access-gqx2z\") pod \"redhat-marketplace-ncgzm\" (UID: \"9f084f23-241a-45b2-88a9-2ffc29b161b4\") " pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.792444 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f084f23-241a-45b2-88a9-2ffc29b161b4-utilities\") pod \"redhat-marketplace-ncgzm\" (UID: \"9f084f23-241a-45b2-88a9-2ffc29b161b4\") " pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.792559 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f084f23-241a-45b2-88a9-2ffc29b161b4-catalog-content\") pod \"redhat-marketplace-ncgzm\" (UID: \"9f084f23-241a-45b2-88a9-2ffc29b161b4\") " pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.794791 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9vz74"] Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.822305 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqx2z\" (UniqueName: \"kubernetes.io/projected/9f084f23-241a-45b2-88a9-2ffc29b161b4-kube-api-access-gqx2z\") pod \"redhat-marketplace-ncgzm\" (UID: \"9f084f23-241a-45b2-88a9-2ffc29b161b4\") " pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.892334 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhrf4\" (UniqueName: \"kubernetes.io/projected/1aa15892-e136-4c58-aaf2-d62597704254-kube-api-access-nhrf4\") pod \"certified-operators-9vz74\" (UID: \"1aa15892-e136-4c58-aaf2-d62597704254\") " pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.892454 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aa15892-e136-4c58-aaf2-d62597704254-utilities\") pod \"certified-operators-9vz74\" (UID: \"1aa15892-e136-4c58-aaf2-d62597704254\") " pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.892489 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aa15892-e136-4c58-aaf2-d62597704254-catalog-content\") pod \"certified-operators-9vz74\" (UID: \"1aa15892-e136-4c58-aaf2-d62597704254\") " pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.892910 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1aa15892-e136-4c58-aaf2-d62597704254-catalog-content\") pod \"certified-operators-9vz74\" (UID: \"1aa15892-e136-4c58-aaf2-d62597704254\") " pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.893126 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1aa15892-e136-4c58-aaf2-d62597704254-utilities\") pod \"certified-operators-9vz74\" (UID: \"1aa15892-e136-4c58-aaf2-d62597704254\") " pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.894420 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:21 crc kubenswrapper[4575]: I1004 04:40:21.910136 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhrf4\" (UniqueName: \"kubernetes.io/projected/1aa15892-e136-4c58-aaf2-d62597704254-kube-api-access-nhrf4\") pod \"certified-operators-9vz74\" (UID: \"1aa15892-e136-4c58-aaf2-d62597704254\") " pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:22 crc kubenswrapper[4575]: I1004 04:40:22.103181 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:22 crc kubenswrapper[4575]: I1004 04:40:22.354017 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9vz74"] Oct 04 04:40:22 crc kubenswrapper[4575]: W1004 04:40:22.363778 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1aa15892_e136_4c58_aaf2_d62597704254.slice/crio-2a75b6b0d9dd1ff2e514e536445fd49504e9b1c98253ffa03db1cfa07689889a WatchSource:0}: Error finding container 2a75b6b0d9dd1ff2e514e536445fd49504e9b1c98253ffa03db1cfa07689889a: Status 404 returned error can't find the container with id 2a75b6b0d9dd1ff2e514e536445fd49504e9b1c98253ffa03db1cfa07689889a Oct 04 04:40:22 crc kubenswrapper[4575]: I1004 04:40:22.429554 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ncgzm"] Oct 04 04:40:22 crc kubenswrapper[4575]: W1004 04:40:22.433690 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f084f23_241a_45b2_88a9_2ffc29b161b4.slice/crio-68d56a63b84e28d539366839822c674c6e1d0a8ad583f540ef7ad100d2da2620 WatchSource:0}: Error finding container 68d56a63b84e28d539366839822c674c6e1d0a8ad583f540ef7ad100d2da2620: Status 404 returned error can't find the container with id 68d56a63b84e28d539366839822c674c6e1d0a8ad583f540ef7ad100d2da2620 Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.000835 4575 generic.go:334] "Generic (PLEG): container finished" podID="1aa15892-e136-4c58-aaf2-d62597704254" containerID="9d06c78f847ca1e988f688ce8cf00aa4d7bb5da4060293dc85cd590003c51eca" exitCode=0 Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.001054 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vz74" event={"ID":"1aa15892-e136-4c58-aaf2-d62597704254","Type":"ContainerDied","Data":"9d06c78f847ca1e988f688ce8cf00aa4d7bb5da4060293dc85cd590003c51eca"} Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.001128 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vz74" event={"ID":"1aa15892-e136-4c58-aaf2-d62597704254","Type":"ContainerStarted","Data":"2a75b6b0d9dd1ff2e514e536445fd49504e9b1c98253ffa03db1cfa07689889a"} Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.004006 4575 generic.go:334] "Generic (PLEG): container finished" podID="9f084f23-241a-45b2-88a9-2ffc29b161b4" containerID="b9037907bad611de16209ab76b19a8c65a5c35878c3db5887a57c627f23d15b2" exitCode=0 Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.004062 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ncgzm" event={"ID":"9f084f23-241a-45b2-88a9-2ffc29b161b4","Type":"ContainerDied","Data":"b9037907bad611de16209ab76b19a8c65a5c35878c3db5887a57c627f23d15b2"} Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.004084 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ncgzm" event={"ID":"9f084f23-241a-45b2-88a9-2ffc29b161b4","Type":"ContainerStarted","Data":"68d56a63b84e28d539366839822c674c6e1d0a8ad583f540ef7ad100d2da2620"} Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.006918 4575 generic.go:334] "Generic (PLEG): container finished" podID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerID="314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda" exitCode=0 Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.007014 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktd5v" event={"ID":"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354","Type":"ContainerDied","Data":"314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda"} Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.009394 4575 generic.go:334] "Generic (PLEG): container finished" podID="cbaf719e-0879-45d7-a285-42109d7fb978" containerID="d09e1d5d9e960d35584c7ad09f51c9457bccebc2708ce4eee58aa4e81de49c36" exitCode=0 Oct 04 04:40:23 crc kubenswrapper[4575]: I1004 04:40:23.009429 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfp6s" event={"ID":"cbaf719e-0879-45d7-a285-42109d7fb978","Type":"ContainerDied","Data":"d09e1d5d9e960d35584c7ad09f51c9457bccebc2708ce4eee58aa4e81de49c36"} Oct 04 04:40:25 crc kubenswrapper[4575]: I1004 04:40:25.028348 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vz74" event={"ID":"1aa15892-e136-4c58-aaf2-d62597704254","Type":"ContainerStarted","Data":"8ce4004d470489acf8b62a33a86ef9f3b0d8cdf0599ea1a955ce3c32f2fa5143"} Oct 04 04:40:25 crc kubenswrapper[4575]: I1004 04:40:25.030466 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ncgzm" event={"ID":"9f084f23-241a-45b2-88a9-2ffc29b161b4","Type":"ContainerStarted","Data":"fc1e421bd0e39046b776c6fb806217cbd59ff6395627239afa8126ae743e23ab"} Oct 04 04:40:25 crc kubenswrapper[4575]: I1004 04:40:25.040361 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktd5v" event={"ID":"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354","Type":"ContainerStarted","Data":"d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac"} Oct 04 04:40:25 crc kubenswrapper[4575]: I1004 04:40:25.045921 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfp6s" event={"ID":"cbaf719e-0879-45d7-a285-42109d7fb978","Type":"ContainerStarted","Data":"d13d00f6f5d875fdce43f01aa6cd846f664dfcfd06929bb38938367c87026828"} Oct 04 04:40:25 crc kubenswrapper[4575]: I1004 04:40:25.072996 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qfp6s" podStartSLOduration=3.316444061 podStartE2EDuration="6.072981421s" podCreationTimestamp="2025-10-04 04:40:19 +0000 UTC" firstStartedPulling="2025-10-04 04:40:20.991442614 +0000 UTC m=+372.320001428" lastFinishedPulling="2025-10-04 04:40:23.747979974 +0000 UTC m=+375.076538788" observedRunningTime="2025-10-04 04:40:25.072628641 +0000 UTC m=+376.401187475" watchObservedRunningTime="2025-10-04 04:40:25.072981421 +0000 UTC m=+376.401540235" Oct 04 04:40:26 crc kubenswrapper[4575]: I1004 04:40:26.052744 4575 generic.go:334] "Generic (PLEG): container finished" podID="1aa15892-e136-4c58-aaf2-d62597704254" containerID="8ce4004d470489acf8b62a33a86ef9f3b0d8cdf0599ea1a955ce3c32f2fa5143" exitCode=0 Oct 04 04:40:26 crc kubenswrapper[4575]: I1004 04:40:26.052952 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vz74" event={"ID":"1aa15892-e136-4c58-aaf2-d62597704254","Type":"ContainerDied","Data":"8ce4004d470489acf8b62a33a86ef9f3b0d8cdf0599ea1a955ce3c32f2fa5143"} Oct 04 04:40:26 crc kubenswrapper[4575]: I1004 04:40:26.058084 4575 generic.go:334] "Generic (PLEG): container finished" podID="9f084f23-241a-45b2-88a9-2ffc29b161b4" containerID="fc1e421bd0e39046b776c6fb806217cbd59ff6395627239afa8126ae743e23ab" exitCode=0 Oct 04 04:40:26 crc kubenswrapper[4575]: I1004 04:40:26.058270 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ncgzm" event={"ID":"9f084f23-241a-45b2-88a9-2ffc29b161b4","Type":"ContainerDied","Data":"fc1e421bd0e39046b776c6fb806217cbd59ff6395627239afa8126ae743e23ab"} Oct 04 04:40:26 crc kubenswrapper[4575]: I1004 04:40:26.075100 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ktd5v" podStartSLOduration=4.158712716 podStartE2EDuration="7.075079805s" podCreationTimestamp="2025-10-04 04:40:19 +0000 UTC" firstStartedPulling="2025-10-04 04:40:20.991475105 +0000 UTC m=+372.320033919" lastFinishedPulling="2025-10-04 04:40:23.907842194 +0000 UTC m=+375.236401008" observedRunningTime="2025-10-04 04:40:25.089676757 +0000 UTC m=+376.418235591" watchObservedRunningTime="2025-10-04 04:40:26.075079805 +0000 UTC m=+377.403638619" Oct 04 04:40:29 crc kubenswrapper[4575]: I1004 04:40:29.501732 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:29 crc kubenswrapper[4575]: I1004 04:40:29.502395 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:29 crc kubenswrapper[4575]: I1004 04:40:29.539658 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:29 crc kubenswrapper[4575]: I1004 04:40:29.713807 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:29 crc kubenswrapper[4575]: I1004 04:40:29.713895 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:29 crc kubenswrapper[4575]: I1004 04:40:29.751650 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:30 crc kubenswrapper[4575]: I1004 04:40:30.087376 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vz74" event={"ID":"1aa15892-e136-4c58-aaf2-d62597704254","Type":"ContainerStarted","Data":"26c235c4b714a0874c9257528a585fa1170252d0d497c14fb4676c20a823002f"} Oct 04 04:40:30 crc kubenswrapper[4575]: I1004 04:40:30.089509 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ncgzm" event={"ID":"9f084f23-241a-45b2-88a9-2ffc29b161b4","Type":"ContainerStarted","Data":"cdb4cb9b92bf73de287db97863f42e361109c920e3522b67d572d599264e9423"} Oct 04 04:40:30 crc kubenswrapper[4575]: I1004 04:40:30.108241 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9vz74" podStartSLOduration=3.058361006 podStartE2EDuration="9.108226486s" podCreationTimestamp="2025-10-04 04:40:21 +0000 UTC" firstStartedPulling="2025-10-04 04:40:23.002955968 +0000 UTC m=+374.331514782" lastFinishedPulling="2025-10-04 04:40:29.052821448 +0000 UTC m=+380.381380262" observedRunningTime="2025-10-04 04:40:30.10770834 +0000 UTC m=+381.436267154" watchObservedRunningTime="2025-10-04 04:40:30.108226486 +0000 UTC m=+381.436785300" Oct 04 04:40:30 crc kubenswrapper[4575]: I1004 04:40:30.136826 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qfp6s" Oct 04 04:40:30 crc kubenswrapper[4575]: I1004 04:40:30.142807 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ktd5v" Oct 04 04:40:31 crc kubenswrapper[4575]: I1004 04:40:31.113752 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ncgzm" podStartSLOduration=3.372002866 podStartE2EDuration="10.113732231s" podCreationTimestamp="2025-10-04 04:40:21 +0000 UTC" firstStartedPulling="2025-10-04 04:40:23.005623318 +0000 UTC m=+374.334182132" lastFinishedPulling="2025-10-04 04:40:29.747352683 +0000 UTC m=+381.075911497" observedRunningTime="2025-10-04 04:40:31.1130294 +0000 UTC m=+382.441588234" watchObservedRunningTime="2025-10-04 04:40:31.113732231 +0000 UTC m=+382.442291045" Oct 04 04:40:31 crc kubenswrapper[4575]: I1004 04:40:31.894647 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:31 crc kubenswrapper[4575]: I1004 04:40:31.894703 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:31 crc kubenswrapper[4575]: I1004 04:40:31.934755 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:32 crc kubenswrapper[4575]: I1004 04:40:32.104580 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:32 crc kubenswrapper[4575]: I1004 04:40:32.104678 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:32 crc kubenswrapper[4575]: I1004 04:40:32.151495 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:40:38 crc kubenswrapper[4575]: I1004 04:40:38.446397 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:40:38 crc kubenswrapper[4575]: I1004 04:40:38.446959 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:40:41 crc kubenswrapper[4575]: I1004 04:40:41.934295 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ncgzm" Oct 04 04:40:42 crc kubenswrapper[4575]: I1004 04:40:42.137180 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9vz74" Oct 04 04:41:08 crc kubenswrapper[4575]: I1004 04:41:08.446160 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:41:08 crc kubenswrapper[4575]: I1004 04:41:08.446687 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:41:38 crc kubenswrapper[4575]: I1004 04:41:38.446107 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:41:38 crc kubenswrapper[4575]: I1004 04:41:38.446865 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:41:38 crc kubenswrapper[4575]: I1004 04:41:38.446934 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:41:38 crc kubenswrapper[4575]: I1004 04:41:38.447755 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7df48bb006932f578e73ad68f06d44012c4eb713d15727b98e4c310b3bde01ac"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:41:38 crc kubenswrapper[4575]: I1004 04:41:38.447854 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://7df48bb006932f578e73ad68f06d44012c4eb713d15727b98e4c310b3bde01ac" gracePeriod=600 Oct 04 04:41:39 crc kubenswrapper[4575]: I1004 04:41:39.432365 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="7df48bb006932f578e73ad68f06d44012c4eb713d15727b98e4c310b3bde01ac" exitCode=0 Oct 04 04:41:39 crc kubenswrapper[4575]: I1004 04:41:39.432413 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"7df48bb006932f578e73ad68f06d44012c4eb713d15727b98e4c310b3bde01ac"} Oct 04 04:41:39 crc kubenswrapper[4575]: I1004 04:41:39.432822 4575 scope.go:117] "RemoveContainer" containerID="efd15a9488030f812f65d49152a8d424d1b925dc86d9a3a95e78ced38472b2db" Oct 04 04:41:40 crc kubenswrapper[4575]: I1004 04:41:40.439638 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"a31a363eba214bd5be24aa256fc82e8758ebfb0d8fe15168b3bda5e51827dda0"} Oct 04 04:43:09 crc kubenswrapper[4575]: I1004 04:43:09.432811 4575 scope.go:117] "RemoveContainer" containerID="7a921e49a9dafc55a414db61be8942c3d244f9c0754ff932610c9a895d17b23a" Oct 04 04:44:08 crc kubenswrapper[4575]: I1004 04:44:08.446284 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:44:08 crc kubenswrapper[4575]: I1004 04:44:08.446871 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:44:38 crc kubenswrapper[4575]: I1004 04:44:38.447103 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:44:38 crc kubenswrapper[4575]: I1004 04:44:38.447748 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.133949 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9"] Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.135172 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.139949 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.140187 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.143909 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9"] Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.263337 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8zs6\" (UniqueName: \"kubernetes.io/projected/f915159b-2f89-4709-a2f7-447ab38f7eaa-kube-api-access-x8zs6\") pod \"collect-profiles-29325885-sl5k9\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.263395 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f915159b-2f89-4709-a2f7-447ab38f7eaa-secret-volume\") pod \"collect-profiles-29325885-sl5k9\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.263425 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f915159b-2f89-4709-a2f7-447ab38f7eaa-config-volume\") pod \"collect-profiles-29325885-sl5k9\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.365322 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8zs6\" (UniqueName: \"kubernetes.io/projected/f915159b-2f89-4709-a2f7-447ab38f7eaa-kube-api-access-x8zs6\") pod \"collect-profiles-29325885-sl5k9\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.365374 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f915159b-2f89-4709-a2f7-447ab38f7eaa-secret-volume\") pod \"collect-profiles-29325885-sl5k9\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.365406 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f915159b-2f89-4709-a2f7-447ab38f7eaa-config-volume\") pod \"collect-profiles-29325885-sl5k9\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.366442 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f915159b-2f89-4709-a2f7-447ab38f7eaa-config-volume\") pod \"collect-profiles-29325885-sl5k9\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.371025 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f915159b-2f89-4709-a2f7-447ab38f7eaa-secret-volume\") pod \"collect-profiles-29325885-sl5k9\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.383822 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8zs6\" (UniqueName: \"kubernetes.io/projected/f915159b-2f89-4709-a2f7-447ab38f7eaa-kube-api-access-x8zs6\") pod \"collect-profiles-29325885-sl5k9\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.454788 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:00 crc kubenswrapper[4575]: I1004 04:45:00.841138 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9"] Oct 04 04:45:01 crc kubenswrapper[4575]: I1004 04:45:01.442616 4575 generic.go:334] "Generic (PLEG): container finished" podID="f915159b-2f89-4709-a2f7-447ab38f7eaa" containerID="9c8c0044b9785c42c82ed24dfcf93de0c1cc8268aad9a97747cd2243f7d33d34" exitCode=0 Oct 04 04:45:01 crc kubenswrapper[4575]: I1004 04:45:01.442703 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" event={"ID":"f915159b-2f89-4709-a2f7-447ab38f7eaa","Type":"ContainerDied","Data":"9c8c0044b9785c42c82ed24dfcf93de0c1cc8268aad9a97747cd2243f7d33d34"} Oct 04 04:45:01 crc kubenswrapper[4575]: I1004 04:45:01.443198 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" event={"ID":"f915159b-2f89-4709-a2f7-447ab38f7eaa","Type":"ContainerStarted","Data":"7f0dbcded60c22b291fb067574edbfb80f9ac7095d167231d07554d38fabd058"} Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.651456 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.798748 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f915159b-2f89-4709-a2f7-447ab38f7eaa-secret-volume\") pod \"f915159b-2f89-4709-a2f7-447ab38f7eaa\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.798831 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8zs6\" (UniqueName: \"kubernetes.io/projected/f915159b-2f89-4709-a2f7-447ab38f7eaa-kube-api-access-x8zs6\") pod \"f915159b-2f89-4709-a2f7-447ab38f7eaa\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.798927 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f915159b-2f89-4709-a2f7-447ab38f7eaa-config-volume\") pod \"f915159b-2f89-4709-a2f7-447ab38f7eaa\" (UID: \"f915159b-2f89-4709-a2f7-447ab38f7eaa\") " Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.799648 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f915159b-2f89-4709-a2f7-447ab38f7eaa-config-volume" (OuterVolumeSpecName: "config-volume") pod "f915159b-2f89-4709-a2f7-447ab38f7eaa" (UID: "f915159b-2f89-4709-a2f7-447ab38f7eaa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.804740 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f915159b-2f89-4709-a2f7-447ab38f7eaa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f915159b-2f89-4709-a2f7-447ab38f7eaa" (UID: "f915159b-2f89-4709-a2f7-447ab38f7eaa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.804823 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f915159b-2f89-4709-a2f7-447ab38f7eaa-kube-api-access-x8zs6" (OuterVolumeSpecName: "kube-api-access-x8zs6") pod "f915159b-2f89-4709-a2f7-447ab38f7eaa" (UID: "f915159b-2f89-4709-a2f7-447ab38f7eaa"). InnerVolumeSpecName "kube-api-access-x8zs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.899934 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8zs6\" (UniqueName: \"kubernetes.io/projected/f915159b-2f89-4709-a2f7-447ab38f7eaa-kube-api-access-x8zs6\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.899968 4575 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f915159b-2f89-4709-a2f7-447ab38f7eaa-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:02 crc kubenswrapper[4575]: I1004 04:45:02.899977 4575 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f915159b-2f89-4709-a2f7-447ab38f7eaa-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:03 crc kubenswrapper[4575]: I1004 04:45:03.455848 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" event={"ID":"f915159b-2f89-4709-a2f7-447ab38f7eaa","Type":"ContainerDied","Data":"7f0dbcded60c22b291fb067574edbfb80f9ac7095d167231d07554d38fabd058"} Oct 04 04:45:03 crc kubenswrapper[4575]: I1004 04:45:03.455903 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f0dbcded60c22b291fb067574edbfb80f9ac7095d167231d07554d38fabd058" Oct 04 04:45:03 crc kubenswrapper[4575]: I1004 04:45:03.455907 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9" Oct 04 04:45:08 crc kubenswrapper[4575]: I1004 04:45:08.446539 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:45:08 crc kubenswrapper[4575]: I1004 04:45:08.446814 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:45:08 crc kubenswrapper[4575]: I1004 04:45:08.446853 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:45:08 crc kubenswrapper[4575]: I1004 04:45:08.447302 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a31a363eba214bd5be24aa256fc82e8758ebfb0d8fe15168b3bda5e51827dda0"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:45:08 crc kubenswrapper[4575]: I1004 04:45:08.447343 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://a31a363eba214bd5be24aa256fc82e8758ebfb0d8fe15168b3bda5e51827dda0" gracePeriod=600 Oct 04 04:45:09 crc kubenswrapper[4575]: I1004 04:45:09.489150 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="a31a363eba214bd5be24aa256fc82e8758ebfb0d8fe15168b3bda5e51827dda0" exitCode=0 Oct 04 04:45:09 crc kubenswrapper[4575]: I1004 04:45:09.489228 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"a31a363eba214bd5be24aa256fc82e8758ebfb0d8fe15168b3bda5e51827dda0"} Oct 04 04:45:09 crc kubenswrapper[4575]: I1004 04:45:09.489510 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"e6a6d9d656e937f48c2e0c38c03877604da45131d454048196bae4e1a3e60a62"} Oct 04 04:45:09 crc kubenswrapper[4575]: I1004 04:45:09.489537 4575 scope.go:117] "RemoveContainer" containerID="7df48bb006932f578e73ad68f06d44012c4eb713d15727b98e4c310b3bde01ac" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.000473 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flrql"] Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.001271 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" podUID="c72540e8-b2dd-4377-8a17-717fcb2a05a2" containerName="controller-manager" containerID="cri-o://a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f" gracePeriod=30 Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.101663 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds"] Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.102167 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" podUID="021e5871-36e8-48d3-92de-e504a9429d9e" containerName="route-controller-manager" containerID="cri-o://0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562" gracePeriod=30 Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.390462 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.463672 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.478832 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-client-ca\") pod \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.478904 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/021e5871-36e8-48d3-92de-e504a9429d9e-serving-cert\") pod \"021e5871-36e8-48d3-92de-e504a9429d9e\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.478965 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-proxy-ca-bundles\") pod \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.478986 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-config\") pod \"021e5871-36e8-48d3-92de-e504a9429d9e\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.479037 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cdkp\" (UniqueName: \"kubernetes.io/projected/c72540e8-b2dd-4377-8a17-717fcb2a05a2-kube-api-access-8cdkp\") pod \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.479054 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls796\" (UniqueName: \"kubernetes.io/projected/021e5871-36e8-48d3-92de-e504a9429d9e-kube-api-access-ls796\") pod \"021e5871-36e8-48d3-92de-e504a9429d9e\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.479078 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c72540e8-b2dd-4377-8a17-717fcb2a05a2-serving-cert\") pod \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.479129 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-config\") pod \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\" (UID: \"c72540e8-b2dd-4377-8a17-717fcb2a05a2\") " Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.479147 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-client-ca\") pod \"021e5871-36e8-48d3-92de-e504a9429d9e\" (UID: \"021e5871-36e8-48d3-92de-e504a9429d9e\") " Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.480471 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-client-ca" (OuterVolumeSpecName: "client-ca") pod "021e5871-36e8-48d3-92de-e504a9429d9e" (UID: "021e5871-36e8-48d3-92de-e504a9429d9e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.481147 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-config" (OuterVolumeSpecName: "config") pod "021e5871-36e8-48d3-92de-e504a9429d9e" (UID: "021e5871-36e8-48d3-92de-e504a9429d9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.482073 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c72540e8-b2dd-4377-8a17-717fcb2a05a2" (UID: "c72540e8-b2dd-4377-8a17-717fcb2a05a2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.482794 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-config" (OuterVolumeSpecName: "config") pod "c72540e8-b2dd-4377-8a17-717fcb2a05a2" (UID: "c72540e8-b2dd-4377-8a17-717fcb2a05a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.484079 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-client-ca" (OuterVolumeSpecName: "client-ca") pod "c72540e8-b2dd-4377-8a17-717fcb2a05a2" (UID: "c72540e8-b2dd-4377-8a17-717fcb2a05a2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.491063 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/021e5871-36e8-48d3-92de-e504a9429d9e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "021e5871-36e8-48d3-92de-e504a9429d9e" (UID: "021e5871-36e8-48d3-92de-e504a9429d9e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.491175 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c72540e8-b2dd-4377-8a17-717fcb2a05a2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c72540e8-b2dd-4377-8a17-717fcb2a05a2" (UID: "c72540e8-b2dd-4377-8a17-717fcb2a05a2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.492138 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/021e5871-36e8-48d3-92de-e504a9429d9e-kube-api-access-ls796" (OuterVolumeSpecName: "kube-api-access-ls796") pod "021e5871-36e8-48d3-92de-e504a9429d9e" (UID: "021e5871-36e8-48d3-92de-e504a9429d9e"). InnerVolumeSpecName "kube-api-access-ls796". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.507857 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c72540e8-b2dd-4377-8a17-717fcb2a05a2-kube-api-access-8cdkp" (OuterVolumeSpecName: "kube-api-access-8cdkp") pod "c72540e8-b2dd-4377-8a17-717fcb2a05a2" (UID: "c72540e8-b2dd-4377-8a17-717fcb2a05a2"). InnerVolumeSpecName "kube-api-access-8cdkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.580885 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c72540e8-b2dd-4377-8a17-717fcb2a05a2-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.580918 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.580927 4575 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.580950 4575 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.580983 4575 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/021e5871-36e8-48d3-92de-e504a9429d9e-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.580993 4575 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c72540e8-b2dd-4377-8a17-717fcb2a05a2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.581003 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021e5871-36e8-48d3-92de-e504a9429d9e-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.581011 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cdkp\" (UniqueName: \"kubernetes.io/projected/c72540e8-b2dd-4377-8a17-717fcb2a05a2-kube-api-access-8cdkp\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.581019 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls796\" (UniqueName: \"kubernetes.io/projected/021e5871-36e8-48d3-92de-e504a9429d9e-kube-api-access-ls796\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.922707 4575 generic.go:334] "Generic (PLEG): container finished" podID="c72540e8-b2dd-4377-8a17-717fcb2a05a2" containerID="a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f" exitCode=0 Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.922767 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.922786 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" event={"ID":"c72540e8-b2dd-4377-8a17-717fcb2a05a2","Type":"ContainerDied","Data":"a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f"} Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.923550 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-flrql" event={"ID":"c72540e8-b2dd-4377-8a17-717fcb2a05a2","Type":"ContainerDied","Data":"df1f9a2773e11e042ffae27840b39f30c13fa3450c2c1fdb0fa007afd5fe06a1"} Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.923602 4575 scope.go:117] "RemoveContainer" containerID="a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.924993 4575 generic.go:334] "Generic (PLEG): container finished" podID="021e5871-36e8-48d3-92de-e504a9429d9e" containerID="0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562" exitCode=0 Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.925221 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.925257 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" event={"ID":"021e5871-36e8-48d3-92de-e504a9429d9e","Type":"ContainerDied","Data":"0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562"} Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.925706 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds" event={"ID":"021e5871-36e8-48d3-92de-e504a9429d9e","Type":"ContainerDied","Data":"e6695130a87c43b98c1b1e876d378e44592156ecc618cb09241abcdd7b671a06"} Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.937961 4575 scope.go:117] "RemoveContainer" containerID="a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f" Oct 04 04:46:39 crc kubenswrapper[4575]: E1004 04:46:39.939014 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f\": container with ID starting with a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f not found: ID does not exist" containerID="a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.939053 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f"} err="failed to get container status \"a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f\": rpc error: code = NotFound desc = could not find container \"a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f\": container with ID starting with a0abbfc55f5b8f408a561c4b323b0643995e119afe2db842414d966eb7c4052f not found: ID does not exist" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.939078 4575 scope.go:117] "RemoveContainer" containerID="0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.953161 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds"] Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.957234 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t2mds"] Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.966677 4575 scope.go:117] "RemoveContainer" containerID="0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562" Oct 04 04:46:39 crc kubenswrapper[4575]: E1004 04:46:39.967405 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562\": container with ID starting with 0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562 not found: ID does not exist" containerID="0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.967436 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562"} err="failed to get container status \"0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562\": rpc error: code = NotFound desc = could not find container \"0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562\": container with ID starting with 0ae71dbdbad6a5605a50fa28a8be5ec5c2d01fe220af3302f4122d9c1777a562 not found: ID does not exist" Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.969376 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flrql"] Oct 04 04:46:39 crc kubenswrapper[4575]: I1004 04:46:39.972760 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flrql"] Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.469239 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n"] Oct 04 04:46:40 crc kubenswrapper[4575]: E1004 04:46:40.469431 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f915159b-2f89-4709-a2f7-447ab38f7eaa" containerName="collect-profiles" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.469443 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f915159b-2f89-4709-a2f7-447ab38f7eaa" containerName="collect-profiles" Oct 04 04:46:40 crc kubenswrapper[4575]: E1004 04:46:40.469457 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="021e5871-36e8-48d3-92de-e504a9429d9e" containerName="route-controller-manager" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.469463 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="021e5871-36e8-48d3-92de-e504a9429d9e" containerName="route-controller-manager" Oct 04 04:46:40 crc kubenswrapper[4575]: E1004 04:46:40.469473 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c72540e8-b2dd-4377-8a17-717fcb2a05a2" containerName="controller-manager" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.469480 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c72540e8-b2dd-4377-8a17-717fcb2a05a2" containerName="controller-manager" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.469562 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="021e5871-36e8-48d3-92de-e504a9429d9e" containerName="route-controller-manager" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.469577 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c72540e8-b2dd-4377-8a17-717fcb2a05a2" containerName="controller-manager" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.469598 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f915159b-2f89-4709-a2f7-447ab38f7eaa" containerName="collect-profiles" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.469966 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.471566 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d9b989955-q9skl"] Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.472013 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.472853 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.473938 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.474128 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.474258 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.474866 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.475117 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.475235 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.476107 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.476229 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.476497 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.476707 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.477198 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.488396 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d9b989955-q9skl"] Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.489029 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.490550 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n"] Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.591095 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1aed8180-077d-4940-b779-be133c3ab31b-client-ca\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.591155 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41646809-f39b-4912-ad9b-1ceb879e7bdf-serving-cert\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.591181 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w28tg\" (UniqueName: \"kubernetes.io/projected/1aed8180-077d-4940-b779-be133c3ab31b-kube-api-access-w28tg\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.591239 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41646809-f39b-4912-ad9b-1ceb879e7bdf-client-ca\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.591273 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aed8180-077d-4940-b779-be133c3ab31b-config\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.591293 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnqpg\" (UniqueName: \"kubernetes.io/projected/41646809-f39b-4912-ad9b-1ceb879e7bdf-kube-api-access-rnqpg\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.591341 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1aed8180-077d-4940-b779-be133c3ab31b-proxy-ca-bundles\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.591360 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41646809-f39b-4912-ad9b-1ceb879e7bdf-config\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.591410 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1aed8180-077d-4940-b779-be133c3ab31b-serving-cert\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.692217 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1aed8180-077d-4940-b779-be133c3ab31b-client-ca\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.692267 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41646809-f39b-4912-ad9b-1ceb879e7bdf-serving-cert\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.692294 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w28tg\" (UniqueName: \"kubernetes.io/projected/1aed8180-077d-4940-b779-be133c3ab31b-kube-api-access-w28tg\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.692316 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41646809-f39b-4912-ad9b-1ceb879e7bdf-client-ca\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.692346 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aed8180-077d-4940-b779-be133c3ab31b-config\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.692364 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnqpg\" (UniqueName: \"kubernetes.io/projected/41646809-f39b-4912-ad9b-1ceb879e7bdf-kube-api-access-rnqpg\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.692394 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1aed8180-077d-4940-b779-be133c3ab31b-proxy-ca-bundles\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.692416 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41646809-f39b-4912-ad9b-1ceb879e7bdf-config\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.692439 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1aed8180-077d-4940-b779-be133c3ab31b-serving-cert\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.693240 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1aed8180-077d-4940-b779-be133c3ab31b-client-ca\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.693510 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41646809-f39b-4912-ad9b-1ceb879e7bdf-client-ca\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.693844 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41646809-f39b-4912-ad9b-1ceb879e7bdf-config\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.694051 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aed8180-077d-4940-b779-be133c3ab31b-config\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.694186 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1aed8180-077d-4940-b779-be133c3ab31b-proxy-ca-bundles\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.696949 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1aed8180-077d-4940-b779-be133c3ab31b-serving-cert\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.696972 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41646809-f39b-4912-ad9b-1ceb879e7bdf-serving-cert\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.714735 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w28tg\" (UniqueName: \"kubernetes.io/projected/1aed8180-077d-4940-b779-be133c3ab31b-kube-api-access-w28tg\") pod \"controller-manager-5d9b989955-q9skl\" (UID: \"1aed8180-077d-4940-b779-be133c3ab31b\") " pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.716492 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnqpg\" (UniqueName: \"kubernetes.io/projected/41646809-f39b-4912-ad9b-1ceb879e7bdf-kube-api-access-rnqpg\") pod \"route-controller-manager-5fb56bc5f8-ql69n\" (UID: \"41646809-f39b-4912-ad9b-1ceb879e7bdf\") " pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.791620 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:40 crc kubenswrapper[4575]: I1004 04:46:40.805519 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.017291 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n"] Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.051559 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d9b989955-q9skl"] Oct 04 04:46:41 crc kubenswrapper[4575]: W1004 04:46:41.056277 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1aed8180_077d_4940_b779_be133c3ab31b.slice/crio-191c498c0038df7c77973fd1c6d4243b2a36ab59cbe8457f5b09fad3e3e2680c WatchSource:0}: Error finding container 191c498c0038df7c77973fd1c6d4243b2a36ab59cbe8457f5b09fad3e3e2680c: Status 404 returned error can't find the container with id 191c498c0038df7c77973fd1c6d4243b2a36ab59cbe8457f5b09fad3e3e2680c Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.315740 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="021e5871-36e8-48d3-92de-e504a9429d9e" path="/var/lib/kubelet/pods/021e5871-36e8-48d3-92de-e504a9429d9e/volumes" Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.316836 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c72540e8-b2dd-4377-8a17-717fcb2a05a2" path="/var/lib/kubelet/pods/c72540e8-b2dd-4377-8a17-717fcb2a05a2/volumes" Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.938422 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" event={"ID":"41646809-f39b-4912-ad9b-1ceb879e7bdf","Type":"ContainerStarted","Data":"511e43486039c76250c83be6de2f2f048dea59de4707dc971f6e8f0a4f39ca2b"} Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.938471 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" event={"ID":"41646809-f39b-4912-ad9b-1ceb879e7bdf","Type":"ContainerStarted","Data":"7eee898d83a6800d8bf5f0fec84db093b798c183cf10aa49fad9bb388ee31aa3"} Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.938749 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.941241 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" event={"ID":"1aed8180-077d-4940-b779-be133c3ab31b","Type":"ContainerStarted","Data":"0a0259032732e42209851c8ec148f7168018068b6f4019d8cbc4513ff0573265"} Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.941269 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" event={"ID":"1aed8180-077d-4940-b779-be133c3ab31b","Type":"ContainerStarted","Data":"191c498c0038df7c77973fd1c6d4243b2a36ab59cbe8457f5b09fad3e3e2680c"} Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.941462 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.945771 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.947128 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.957082 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5fb56bc5f8-ql69n" podStartSLOduration=2.957046381 podStartE2EDuration="2.957046381s" podCreationTimestamp="2025-10-04 04:46:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:46:41.954365814 +0000 UTC m=+753.282924628" watchObservedRunningTime="2025-10-04 04:46:41.957046381 +0000 UTC m=+753.285605205" Oct 04 04:46:41 crc kubenswrapper[4575]: I1004 04:46:41.969120 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d9b989955-q9skl" podStartSLOduration=2.969103638 podStartE2EDuration="2.969103638s" podCreationTimestamp="2025-10-04 04:46:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:46:41.968113119 +0000 UTC m=+753.296671943" watchObservedRunningTime="2025-10-04 04:46:41.969103638 +0000 UTC m=+753.297662452" Oct 04 04:46:42 crc kubenswrapper[4575]: I1004 04:46:42.664324 4575 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 04:47:08 crc kubenswrapper[4575]: I1004 04:47:08.446520 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:47:08 crc kubenswrapper[4575]: I1004 04:47:08.447043 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:47:38 crc kubenswrapper[4575]: I1004 04:47:38.446057 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:47:38 crc kubenswrapper[4575]: I1004 04:47:38.446634 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.641599 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-d8xt7"] Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.644060 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-d8xt7" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.648845 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.649090 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.649274 4575 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wwj95" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.663911 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9zwj6"] Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.664742 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9zwj6" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.666764 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-d8xt7"] Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.672387 4575 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-whgrf" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.692597 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-j4lff"] Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.693285 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.695556 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9zwj6"] Oct 04 04:47:47 crc kubenswrapper[4575]: W1004 04:47:47.700671 4575 reflector.go:561] object-"cert-manager"/"cert-manager-webhook-dockercfg-m6fk7": failed to list *v1.Secret: secrets "cert-manager-webhook-dockercfg-m6fk7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Oct 04 04:47:47 crc kubenswrapper[4575]: E1004 04:47:47.700707 4575 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"cert-manager-webhook-dockercfg-m6fk7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-manager-webhook-dockercfg-m6fk7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.730968 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-j4lff"] Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.787095 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzz7j\" (UniqueName: \"kubernetes.io/projected/9eb142f2-2b06-40ea-8472-67052416f1e1-kube-api-access-kzz7j\") pod \"cert-manager-webhook-5655c58dd6-j4lff\" (UID: \"9eb142f2-2b06-40ea-8472-67052416f1e1\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.787149 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bm4j\" (UniqueName: \"kubernetes.io/projected/49a0e02e-c4bb-4b20-afa8-3c7cf160cc2f-kube-api-access-5bm4j\") pod \"cert-manager-5b446d88c5-9zwj6\" (UID: \"49a0e02e-c4bb-4b20-afa8-3c7cf160cc2f\") " pod="cert-manager/cert-manager-5b446d88c5-9zwj6" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.787175 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr42j\" (UniqueName: \"kubernetes.io/projected/84f5ce18-246c-4e87-83df-c7e31b1d71ec-kube-api-access-xr42j\") pod \"cert-manager-cainjector-7f985d654d-d8xt7\" (UID: \"84f5ce18-246c-4e87-83df-c7e31b1d71ec\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-d8xt7" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.888891 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzz7j\" (UniqueName: \"kubernetes.io/projected/9eb142f2-2b06-40ea-8472-67052416f1e1-kube-api-access-kzz7j\") pod \"cert-manager-webhook-5655c58dd6-j4lff\" (UID: \"9eb142f2-2b06-40ea-8472-67052416f1e1\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.889267 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bm4j\" (UniqueName: \"kubernetes.io/projected/49a0e02e-c4bb-4b20-afa8-3c7cf160cc2f-kube-api-access-5bm4j\") pod \"cert-manager-5b446d88c5-9zwj6\" (UID: \"49a0e02e-c4bb-4b20-afa8-3c7cf160cc2f\") " pod="cert-manager/cert-manager-5b446d88c5-9zwj6" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.889387 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr42j\" (UniqueName: \"kubernetes.io/projected/84f5ce18-246c-4e87-83df-c7e31b1d71ec-kube-api-access-xr42j\") pod \"cert-manager-cainjector-7f985d654d-d8xt7\" (UID: \"84f5ce18-246c-4e87-83df-c7e31b1d71ec\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-d8xt7" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.912553 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzz7j\" (UniqueName: \"kubernetes.io/projected/9eb142f2-2b06-40ea-8472-67052416f1e1-kube-api-access-kzz7j\") pod \"cert-manager-webhook-5655c58dd6-j4lff\" (UID: \"9eb142f2-2b06-40ea-8472-67052416f1e1\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.912651 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bm4j\" (UniqueName: \"kubernetes.io/projected/49a0e02e-c4bb-4b20-afa8-3c7cf160cc2f-kube-api-access-5bm4j\") pod \"cert-manager-5b446d88c5-9zwj6\" (UID: \"49a0e02e-c4bb-4b20-afa8-3c7cf160cc2f\") " pod="cert-manager/cert-manager-5b446d88c5-9zwj6" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.918753 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr42j\" (UniqueName: \"kubernetes.io/projected/84f5ce18-246c-4e87-83df-c7e31b1d71ec-kube-api-access-xr42j\") pod \"cert-manager-cainjector-7f985d654d-d8xt7\" (UID: \"84f5ce18-246c-4e87-83df-c7e31b1d71ec\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-d8xt7" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.963337 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-d8xt7" Oct 04 04:47:47 crc kubenswrapper[4575]: I1004 04:47:47.978943 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9zwj6" Oct 04 04:47:48 crc kubenswrapper[4575]: I1004 04:47:48.383896 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-d8xt7"] Oct 04 04:47:48 crc kubenswrapper[4575]: I1004 04:47:48.396198 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:47:48 crc kubenswrapper[4575]: I1004 04:47:48.431299 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9zwj6"] Oct 04 04:47:48 crc kubenswrapper[4575]: W1004 04:47:48.433152 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a0e02e_c4bb_4b20_afa8_3c7cf160cc2f.slice/crio-19f7f6563b44b0cabb6e6b3f9249da8a6068cc3fa92550e9f6fa687fa827a929 WatchSource:0}: Error finding container 19f7f6563b44b0cabb6e6b3f9249da8a6068cc3fa92550e9f6fa687fa827a929: Status 404 returned error can't find the container with id 19f7f6563b44b0cabb6e6b3f9249da8a6068cc3fa92550e9f6fa687fa827a929 Oct 04 04:47:49 crc kubenswrapper[4575]: I1004 04:47:49.009815 4575 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 04 04:47:49 crc kubenswrapper[4575]: I1004 04:47:49.009908 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" Oct 04 04:47:49 crc kubenswrapper[4575]: I1004 04:47:49.078784 4575 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-m6fk7" Oct 04 04:47:49 crc kubenswrapper[4575]: I1004 04:47:49.259910 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9zwj6" event={"ID":"49a0e02e-c4bb-4b20-afa8-3c7cf160cc2f","Type":"ContainerStarted","Data":"19f7f6563b44b0cabb6e6b3f9249da8a6068cc3fa92550e9f6fa687fa827a929"} Oct 04 04:47:49 crc kubenswrapper[4575]: I1004 04:47:49.261168 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-d8xt7" event={"ID":"84f5ce18-246c-4e87-83df-c7e31b1d71ec","Type":"ContainerStarted","Data":"bfeb27ee12621b28ff921231146f84534f3330a814ecce790099cd99acca40b4"} Oct 04 04:47:49 crc kubenswrapper[4575]: I1004 04:47:49.407790 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-j4lff"] Oct 04 04:47:50 crc kubenswrapper[4575]: I1004 04:47:50.266216 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" event={"ID":"9eb142f2-2b06-40ea-8472-67052416f1e1","Type":"ContainerStarted","Data":"9af167328f06b0cce96d377d254f0a6a2e1b40763033c03373ded6c89d815581"} Oct 04 04:47:51 crc kubenswrapper[4575]: I1004 04:47:51.272522 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9zwj6" event={"ID":"49a0e02e-c4bb-4b20-afa8-3c7cf160cc2f","Type":"ContainerStarted","Data":"b826221765a5b18fdc64ee3ebc47f1b2df506f7edd7ae8106bcc5f1cafcf3c78"} Oct 04 04:47:51 crc kubenswrapper[4575]: I1004 04:47:51.288676 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-9zwj6" podStartSLOduration=1.748522814 podStartE2EDuration="4.288661603s" podCreationTimestamp="2025-10-04 04:47:47 +0000 UTC" firstStartedPulling="2025-10-04 04:47:48.435777494 +0000 UTC m=+819.764336308" lastFinishedPulling="2025-10-04 04:47:50.975916283 +0000 UTC m=+822.304475097" observedRunningTime="2025-10-04 04:47:51.287254433 +0000 UTC m=+822.615813247" watchObservedRunningTime="2025-10-04 04:47:51.288661603 +0000 UTC m=+822.617220417" Oct 04 04:47:53 crc kubenswrapper[4575]: I1004 04:47:53.282713 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" event={"ID":"9eb142f2-2b06-40ea-8472-67052416f1e1","Type":"ContainerStarted","Data":"0f5156bd12037f5830807114259418e81f9415ca21d74622422701d05de238af"} Oct 04 04:47:53 crc kubenswrapper[4575]: I1004 04:47:53.283693 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" Oct 04 04:47:53 crc kubenswrapper[4575]: I1004 04:47:53.284206 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-d8xt7" event={"ID":"84f5ce18-246c-4e87-83df-c7e31b1d71ec","Type":"ContainerStarted","Data":"a5bdb855051391b1a9e3e2c87ea1ec8893efb45bdcc4875975a71b3bc0eaacfc"} Oct 04 04:47:53 crc kubenswrapper[4575]: I1004 04:47:53.297000 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" podStartSLOduration=3.302006111 podStartE2EDuration="6.296981795s" podCreationTimestamp="2025-10-04 04:47:47 +0000 UTC" firstStartedPulling="2025-10-04 04:47:49.415035293 +0000 UTC m=+820.743594107" lastFinishedPulling="2025-10-04 04:47:52.410010977 +0000 UTC m=+823.738569791" observedRunningTime="2025-10-04 04:47:53.295600915 +0000 UTC m=+824.624159729" watchObservedRunningTime="2025-10-04 04:47:53.296981795 +0000 UTC m=+824.625540609" Oct 04 04:47:53 crc kubenswrapper[4575]: I1004 04:47:53.309352 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-d8xt7" podStartSLOduration=2.361045353 podStartE2EDuration="6.30933406s" podCreationTimestamp="2025-10-04 04:47:47 +0000 UTC" firstStartedPulling="2025-10-04 04:47:48.395943299 +0000 UTC m=+819.724502103" lastFinishedPulling="2025-10-04 04:47:52.344232006 +0000 UTC m=+823.672790810" observedRunningTime="2025-10-04 04:47:53.306497208 +0000 UTC m=+824.635056032" watchObservedRunningTime="2025-10-04 04:47:53.30933406 +0000 UTC m=+824.637892874" Oct 04 04:47:57 crc kubenswrapper[4575]: I1004 04:47:57.622645 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hqg47"] Oct 04 04:47:57 crc kubenswrapper[4575]: I1004 04:47:57.623271 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovn-controller" containerID="cri-o://bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d" gracePeriod=30 Oct 04 04:47:57 crc kubenswrapper[4575]: I1004 04:47:57.623399 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="northd" containerID="cri-o://f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136" gracePeriod=30 Oct 04 04:47:57 crc kubenswrapper[4575]: I1004 04:47:57.623435 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d" gracePeriod=30 Oct 04 04:47:57 crc kubenswrapper[4575]: I1004 04:47:57.623414 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="nbdb" containerID="cri-o://23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf" gracePeriod=30 Oct 04 04:47:57 crc kubenswrapper[4575]: I1004 04:47:57.623471 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kube-rbac-proxy-node" containerID="cri-o://e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077" gracePeriod=30 Oct 04 04:47:57 crc kubenswrapper[4575]: I1004 04:47:57.623481 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovn-acl-logging" containerID="cri-o://636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e" gracePeriod=30 Oct 04 04:47:57 crc kubenswrapper[4575]: I1004 04:47:57.623750 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="sbdb" containerID="cri-o://dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d" gracePeriod=30 Oct 04 04:47:57 crc kubenswrapper[4575]: I1004 04:47:57.685719 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" containerID="cri-o://028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310" gracePeriod=30 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.033546 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/3.log" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.035458 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovn-acl-logging/0.log" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.035926 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovn-controller/0.log" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.036356 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089331 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jd8rt"] Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089669 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089693 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089704 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089714 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089729 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089739 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089747 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089754 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089764 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="sbdb" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089771 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="sbdb" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089779 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kubecfg-setup" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089785 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kubecfg-setup" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089804 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovn-acl-logging" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089811 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovn-acl-logging" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089824 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovn-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089832 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovn-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089841 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="nbdb" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089848 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="nbdb" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089859 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089867 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089879 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="northd" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089890 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="northd" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.089902 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kube-rbac-proxy-node" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.089912 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kube-rbac-proxy-node" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090065 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="nbdb" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090086 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovn-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090098 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="sbdb" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090109 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090115 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090123 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090129 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kube-rbac-proxy-node" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090139 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="northd" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090147 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovn-acl-logging" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090154 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.090244 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090252 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090357 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.090559 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="984bf56e-d83a-4206-a887-116d728163d1" containerName="ovnkube-controller" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.092929 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122175 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-kubelet\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122219 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-openvswitch\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122247 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-ovn\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122280 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfjxd\" (UniqueName: \"kubernetes.io/projected/984bf56e-d83a-4206-a887-116d728163d1-kube-api-access-sfjxd\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122303 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-var-lib-openvswitch\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122329 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-systemd-units\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122396 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122387 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122409 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122426 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122349 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-etc-openvswitch\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122460 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122544 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-node-log\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122615 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-node-log" (OuterVolumeSpecName: "node-log") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122634 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-log-socket\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122653 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-log-socket" (OuterVolumeSpecName: "log-socket") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122673 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-systemd\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122732 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122765 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-script-lib\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122792 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-env-overrides\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122836 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-bin\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122872 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-ovn-kubernetes\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122899 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-slash\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122891 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122920 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-netns\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122947 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.122970 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-netd\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123000 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-config\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123042 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/984bf56e-d83a-4206-a887-116d728163d1-ovn-node-metrics-cert\") pod \"984bf56e-d83a-4206-a887-116d728163d1\" (UID: \"984bf56e-d83a-4206-a887-116d728163d1\") " Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123297 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-run-netns\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123332 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-log-socket\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123347 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123359 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-run-ovn-kubernetes\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123364 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123384 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123392 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-cni-netd\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123405 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-slash" (OuterVolumeSpecName: "host-slash") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123409 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123425 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123534 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-slash\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123614 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-run-systemd\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123681 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-kubelet\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123697 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-node-log\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123675 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123749 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123820 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvqr9\" (UniqueName: \"kubernetes.io/projected/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-kube-api-access-jvqr9\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123835 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123908 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-ovnkube-config\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.123930 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-var-lib-openvswitch\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124019 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-run-openvswitch\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124036 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-ovnkube-script-lib\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124057 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-systemd-units\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124097 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-env-overrides\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124138 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-run-ovn\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124154 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-ovn-node-metrics-cert\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124194 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-cni-bin\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124218 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-etc-openvswitch\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124316 4575 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124328 4575 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124338 4575 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-slash\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124346 4575 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124354 4575 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124362 4575 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124373 4575 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124381 4575 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124391 4575 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124402 4575 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124412 4575 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124421 4575 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124429 4575 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-node-log\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124442 4575 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-log-socket\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124453 4575 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124462 4575 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.124472 4575 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/984bf56e-d83a-4206-a887-116d728163d1-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.129282 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/984bf56e-d83a-4206-a887-116d728163d1-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.130801 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/984bf56e-d83a-4206-a887-116d728163d1-kube-api-access-sfjxd" (OuterVolumeSpecName: "kube-api-access-sfjxd") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "kube-api-access-sfjxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.140132 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "984bf56e-d83a-4206-a887-116d728163d1" (UID: "984bf56e-d83a-4206-a887-116d728163d1"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.225898 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-run-netns\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.225945 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-log-socket\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.225975 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-run-ovn-kubernetes\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.225991 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-cni-netd\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226020 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-slash\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226045 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-run-systemd\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226062 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-kubelet\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226074 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-node-log\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226092 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226114 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvqr9\" (UniqueName: \"kubernetes.io/projected/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-kube-api-access-jvqr9\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226096 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-run-netns\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226133 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-ovnkube-config\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226247 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-var-lib-openvswitch\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226359 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-systemd-units\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226385 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-run-openvswitch\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226412 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-ovnkube-script-lib\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226473 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-env-overrides\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226508 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-run-ovn\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226535 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-ovn-node-metrics-cert\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226686 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-cni-bin\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226713 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-etc-openvswitch\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226803 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-ovnkube-config\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226859 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-cni-netd\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226872 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-run-ovn-kubernetes\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226882 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-slash\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226912 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-var-lib-openvswitch\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226946 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-systemd-units\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.226972 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-run-openvswitch\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227425 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-ovnkube-script-lib\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227654 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227734 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-etc-openvswitch\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227766 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-env-overrides\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227661 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-run-systemd\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227710 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-cni-bin\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227821 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-host-kubelet\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227850 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-node-log\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227875 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-log-socket\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227892 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-run-ovn\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.227986 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfjxd\" (UniqueName: \"kubernetes.io/projected/984bf56e-d83a-4206-a887-116d728163d1-kube-api-access-sfjxd\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.228003 4575 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/984bf56e-d83a-4206-a887-116d728163d1-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.228014 4575 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/984bf56e-d83a-4206-a887-116d728163d1-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.231185 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-ovn-node-metrics-cert\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.244105 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvqr9\" (UniqueName: \"kubernetes.io/projected/94b8f7bb-4568-4d39-a49e-7df9e84ae94d-kube-api-access-jvqr9\") pod \"ovnkube-node-jd8rt\" (UID: \"94b8f7bb-4568-4d39-a49e-7df9e84ae94d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.309800 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sj5hq_31197ccb-a3f9-4130-bca4-ca1c3d3ee065/kube-multus/2.log" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.310390 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sj5hq_31197ccb-a3f9-4130-bca4-ca1c3d3ee065/kube-multus/1.log" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.310449 4575 generic.go:334] "Generic (PLEG): container finished" podID="31197ccb-a3f9-4130-bca4-ca1c3d3ee065" containerID="c3913b32c9fe101cb5dd4815993d5e8747107d47f003f32bdcade9542f016a66" exitCode=2 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.310520 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sj5hq" event={"ID":"31197ccb-a3f9-4130-bca4-ca1c3d3ee065","Type":"ContainerDied","Data":"c3913b32c9fe101cb5dd4815993d5e8747107d47f003f32bdcade9542f016a66"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.310561 4575 scope.go:117] "RemoveContainer" containerID="ca0e6db6b851d918ef6f7dc44fbfddb61440fb6664c2dfbef5d208c513c3d57f" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.311138 4575 scope.go:117] "RemoveContainer" containerID="c3913b32c9fe101cb5dd4815993d5e8747107d47f003f32bdcade9542f016a66" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.315457 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovnkube-controller/3.log" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.320097 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovn-acl-logging/0.log" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.320827 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hqg47_984bf56e-d83a-4206-a887-116d728163d1/ovn-controller/0.log" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321222 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310" exitCode=0 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321251 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d" exitCode=0 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321261 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf" exitCode=0 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321268 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136" exitCode=0 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321275 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d" exitCode=0 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321280 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077" exitCode=0 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321302 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e" exitCode=143 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321314 4575 generic.go:334] "Generic (PLEG): container finished" podID="984bf56e-d83a-4206-a887-116d728163d1" containerID="bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d" exitCode=143 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321336 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321341 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321365 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321378 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321387 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321397 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321409 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321419 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321428 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321433 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321438 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321443 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321449 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321453 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321460 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321466 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321471 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321478 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321502 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321510 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321515 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321521 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321526 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321531 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321536 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321541 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321545 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321550 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321560 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321567 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321574 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321599 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321606 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321611 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321616 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321622 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321628 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321634 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321639 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321646 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqg47" event={"ID":"984bf56e-d83a-4206-a887-116d728163d1","Type":"ContainerDied","Data":"b58883a55552790c35d5aef956599fc86b2461ad141bc0859dabdde482938a83"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321653 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321659 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321664 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321669 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321674 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321679 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321684 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321691 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321697 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.321704 4575 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.369677 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hqg47"] Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.374202 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hqg47"] Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.407134 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.419229 4575 scope.go:117] "RemoveContainer" containerID="028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310" Oct 04 04:47:58 crc kubenswrapper[4575]: W1004 04:47:58.429196 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94b8f7bb_4568_4d39_a49e_7df9e84ae94d.slice/crio-d60062add529128c0622c258a1adb9a21e5713f58ba5f0bb6720da1a137725c2 WatchSource:0}: Error finding container d60062add529128c0622c258a1adb9a21e5713f58ba5f0bb6720da1a137725c2: Status 404 returned error can't find the container with id d60062add529128c0622c258a1adb9a21e5713f58ba5f0bb6720da1a137725c2 Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.449067 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.485990 4575 scope.go:117] "RemoveContainer" containerID="dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.519763 4575 scope.go:117] "RemoveContainer" containerID="23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.534047 4575 scope.go:117] "RemoveContainer" containerID="f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.552530 4575 scope.go:117] "RemoveContainer" containerID="674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.566808 4575 scope.go:117] "RemoveContainer" containerID="e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.640065 4575 scope.go:117] "RemoveContainer" containerID="636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.653900 4575 scope.go:117] "RemoveContainer" containerID="bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.670064 4575 scope.go:117] "RemoveContainer" containerID="260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.684549 4575 scope.go:117] "RemoveContainer" containerID="028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.685155 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310\": container with ID starting with 028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310 not found: ID does not exist" containerID="028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.685230 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} err="failed to get container status \"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310\": rpc error: code = NotFound desc = could not find container \"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310\": container with ID starting with 028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.685286 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.685679 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\": container with ID starting with 483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83 not found: ID does not exist" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.685710 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} err="failed to get container status \"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\": rpc error: code = NotFound desc = could not find container \"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\": container with ID starting with 483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.685728 4575 scope.go:117] "RemoveContainer" containerID="dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.686080 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\": container with ID starting with dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d not found: ID does not exist" containerID="dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.686108 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} err="failed to get container status \"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\": rpc error: code = NotFound desc = could not find container \"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\": container with ID starting with dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.686124 4575 scope.go:117] "RemoveContainer" containerID="23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.686569 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\": container with ID starting with 23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf not found: ID does not exist" containerID="23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.686718 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} err="failed to get container status \"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\": rpc error: code = NotFound desc = could not find container \"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\": container with ID starting with 23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.686763 4575 scope.go:117] "RemoveContainer" containerID="f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.687217 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\": container with ID starting with f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136 not found: ID does not exist" containerID="f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.687279 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} err="failed to get container status \"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\": rpc error: code = NotFound desc = could not find container \"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\": container with ID starting with f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.687319 4575 scope.go:117] "RemoveContainer" containerID="674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.687638 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\": container with ID starting with 674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d not found: ID does not exist" containerID="674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.687672 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} err="failed to get container status \"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\": rpc error: code = NotFound desc = could not find container \"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\": container with ID starting with 674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.687690 4575 scope.go:117] "RemoveContainer" containerID="e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.688052 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\": container with ID starting with e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077 not found: ID does not exist" containerID="e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.688084 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} err="failed to get container status \"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\": rpc error: code = NotFound desc = could not find container \"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\": container with ID starting with e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.688127 4575 scope.go:117] "RemoveContainer" containerID="636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.688478 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\": container with ID starting with 636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e not found: ID does not exist" containerID="636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.688512 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} err="failed to get container status \"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\": rpc error: code = NotFound desc = could not find container \"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\": container with ID starting with 636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.688530 4575 scope.go:117] "RemoveContainer" containerID="bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.688841 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\": container with ID starting with bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d not found: ID does not exist" containerID="bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.688870 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} err="failed to get container status \"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\": rpc error: code = NotFound desc = could not find container \"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\": container with ID starting with bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.688885 4575 scope.go:117] "RemoveContainer" containerID="260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f" Oct 04 04:47:58 crc kubenswrapper[4575]: E1004 04:47:58.689191 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\": container with ID starting with 260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f not found: ID does not exist" containerID="260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.689227 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} err="failed to get container status \"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\": rpc error: code = NotFound desc = could not find container \"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\": container with ID starting with 260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.689246 4575 scope.go:117] "RemoveContainer" containerID="028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.689550 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} err="failed to get container status \"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310\": rpc error: code = NotFound desc = could not find container \"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310\": container with ID starting with 028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.689573 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.689921 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} err="failed to get container status \"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\": rpc error: code = NotFound desc = could not find container \"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\": container with ID starting with 483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.689950 4575 scope.go:117] "RemoveContainer" containerID="dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.690276 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} err="failed to get container status \"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\": rpc error: code = NotFound desc = could not find container \"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\": container with ID starting with dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.690301 4575 scope.go:117] "RemoveContainer" containerID="23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.690696 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} err="failed to get container status \"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\": rpc error: code = NotFound desc = could not find container \"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\": container with ID starting with 23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.691993 4575 scope.go:117] "RemoveContainer" containerID="f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.692382 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} err="failed to get container status \"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\": rpc error: code = NotFound desc = could not find container \"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\": container with ID starting with f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.692408 4575 scope.go:117] "RemoveContainer" containerID="674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.694343 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} err="failed to get container status \"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\": rpc error: code = NotFound desc = could not find container \"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\": container with ID starting with 674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.694379 4575 scope.go:117] "RemoveContainer" containerID="e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.694711 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} err="failed to get container status \"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\": rpc error: code = NotFound desc = could not find container \"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\": container with ID starting with e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.694757 4575 scope.go:117] "RemoveContainer" containerID="636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.695067 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} err="failed to get container status \"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\": rpc error: code = NotFound desc = could not find container \"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\": container with ID starting with 636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.695094 4575 scope.go:117] "RemoveContainer" containerID="bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.695371 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} err="failed to get container status \"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\": rpc error: code = NotFound desc = could not find container \"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\": container with ID starting with bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.695396 4575 scope.go:117] "RemoveContainer" containerID="260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.695667 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} err="failed to get container status \"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\": rpc error: code = NotFound desc = could not find container \"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\": container with ID starting with 260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.695694 4575 scope.go:117] "RemoveContainer" containerID="028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.695996 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} err="failed to get container status \"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310\": rpc error: code = NotFound desc = could not find container \"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310\": container with ID starting with 028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.696023 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.696259 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} err="failed to get container status \"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\": rpc error: code = NotFound desc = could not find container \"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\": container with ID starting with 483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.696282 4575 scope.go:117] "RemoveContainer" containerID="dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.696514 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} err="failed to get container status \"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\": rpc error: code = NotFound desc = could not find container \"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\": container with ID starting with dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.696535 4575 scope.go:117] "RemoveContainer" containerID="23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.696804 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} err="failed to get container status \"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\": rpc error: code = NotFound desc = could not find container \"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\": container with ID starting with 23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.696838 4575 scope.go:117] "RemoveContainer" containerID="f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.697107 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} err="failed to get container status \"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\": rpc error: code = NotFound desc = could not find container \"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\": container with ID starting with f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.697139 4575 scope.go:117] "RemoveContainer" containerID="674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.697447 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} err="failed to get container status \"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\": rpc error: code = NotFound desc = could not find container \"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\": container with ID starting with 674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.697470 4575 scope.go:117] "RemoveContainer" containerID="e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.697737 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} err="failed to get container status \"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\": rpc error: code = NotFound desc = could not find container \"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\": container with ID starting with e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.697763 4575 scope.go:117] "RemoveContainer" containerID="636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.698097 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} err="failed to get container status \"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\": rpc error: code = NotFound desc = could not find container \"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\": container with ID starting with 636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.698120 4575 scope.go:117] "RemoveContainer" containerID="bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.698406 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} err="failed to get container status \"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\": rpc error: code = NotFound desc = could not find container \"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\": container with ID starting with bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.698431 4575 scope.go:117] "RemoveContainer" containerID="260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.698712 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} err="failed to get container status \"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\": rpc error: code = NotFound desc = could not find container \"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\": container with ID starting with 260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.698739 4575 scope.go:117] "RemoveContainer" containerID="028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.699002 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310"} err="failed to get container status \"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310\": rpc error: code = NotFound desc = could not find container \"028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310\": container with ID starting with 028c582b99556bd480381c35da7b43c963e736e52723a4328cb17f704332d310 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.699033 4575 scope.go:117] "RemoveContainer" containerID="483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.699297 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83"} err="failed to get container status \"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\": rpc error: code = NotFound desc = could not find container \"483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83\": container with ID starting with 483e3621dcd956f99002f0c5174c71ceff0d80efb2b2217621cdf0dbfcbc4e83 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.699318 4575 scope.go:117] "RemoveContainer" containerID="dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.699542 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d"} err="failed to get container status \"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\": rpc error: code = NotFound desc = could not find container \"dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d\": container with ID starting with dc90784a1ed290529326179d8fae07d12bd5c0055e7a5b8a6984bdd94e882c3d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.699566 4575 scope.go:117] "RemoveContainer" containerID="23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.700045 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf"} err="failed to get container status \"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\": rpc error: code = NotFound desc = could not find container \"23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf\": container with ID starting with 23bcd953969edc8dc57ca4529896a714bd5a7b22e586f8f2bcc0663aaf29dbdf not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.700071 4575 scope.go:117] "RemoveContainer" containerID="f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.700346 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136"} err="failed to get container status \"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\": rpc error: code = NotFound desc = could not find container \"f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136\": container with ID starting with f6620c7ab9d5a4bc447b24981cba0f816135b7284eaa3c1a9b4b928503435136 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.700384 4575 scope.go:117] "RemoveContainer" containerID="674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.700679 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d"} err="failed to get container status \"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\": rpc error: code = NotFound desc = could not find container \"674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d\": container with ID starting with 674f9930975612565fc8498eff4acbe8053e54a0e273c4c70713b096caeaff4d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.700700 4575 scope.go:117] "RemoveContainer" containerID="e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.700941 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077"} err="failed to get container status \"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\": rpc error: code = NotFound desc = could not find container \"e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077\": container with ID starting with e627f2e7712bba9e0943037f0ac7809a0ab464e598b7fb57f90a51d94b4df077 not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.700973 4575 scope.go:117] "RemoveContainer" containerID="636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.701257 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e"} err="failed to get container status \"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\": rpc error: code = NotFound desc = could not find container \"636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e\": container with ID starting with 636beb779ea016215357a3b9967328ae1b7c1ed973e3b0b0ca6cbfd7b1fd1e6e not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.701281 4575 scope.go:117] "RemoveContainer" containerID="bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.701525 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d"} err="failed to get container status \"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\": rpc error: code = NotFound desc = could not find container \"bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d\": container with ID starting with bee1920aee2f93a3ab3f3072b557f0f67caf6fd2732ff00251211d43deb53e0d not found: ID does not exist" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.701550 4575 scope.go:117] "RemoveContainer" containerID="260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f" Oct 04 04:47:58 crc kubenswrapper[4575]: I1004 04:47:58.701837 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f"} err="failed to get container status \"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\": rpc error: code = NotFound desc = could not find container \"260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f\": container with ID starting with 260527d7fba961777fcc0d83224d24316e8b1033879fe22d5c8bfc1061b4308f not found: ID does not exist" Oct 04 04:47:59 crc kubenswrapper[4575]: I1004 04:47:59.012611 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4lff" Oct 04 04:47:59 crc kubenswrapper[4575]: I1004 04:47:59.318577 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="984bf56e-d83a-4206-a887-116d728163d1" path="/var/lib/kubelet/pods/984bf56e-d83a-4206-a887-116d728163d1/volumes" Oct 04 04:47:59 crc kubenswrapper[4575]: I1004 04:47:59.328795 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sj5hq_31197ccb-a3f9-4130-bca4-ca1c3d3ee065/kube-multus/2.log" Oct 04 04:47:59 crc kubenswrapper[4575]: I1004 04:47:59.328904 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sj5hq" event={"ID":"31197ccb-a3f9-4130-bca4-ca1c3d3ee065","Type":"ContainerStarted","Data":"b37afa4a48e3736ccb312eb32fef81cbd120d5618dbb0a7f013d7f026fccfd27"} Oct 04 04:47:59 crc kubenswrapper[4575]: I1004 04:47:59.332907 4575 generic.go:334] "Generic (PLEG): container finished" podID="94b8f7bb-4568-4d39-a49e-7df9e84ae94d" containerID="bcc62d3c32d6bc5337900d22a293603033b202efead35756cb86be1f52e73c0b" exitCode=0 Oct 04 04:47:59 crc kubenswrapper[4575]: I1004 04:47:59.332975 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerDied","Data":"bcc62d3c32d6bc5337900d22a293603033b202efead35756cb86be1f52e73c0b"} Oct 04 04:47:59 crc kubenswrapper[4575]: I1004 04:47:59.333038 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerStarted","Data":"d60062add529128c0622c258a1adb9a21e5713f58ba5f0bb6720da1a137725c2"} Oct 04 04:48:00 crc kubenswrapper[4575]: I1004 04:48:00.342249 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerStarted","Data":"d32680fb27bc66edc771ea8786088aed03f42d50f0a72c5278edf2cb624506ae"} Oct 04 04:48:00 crc kubenswrapper[4575]: I1004 04:48:00.342670 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerStarted","Data":"a5e9217961d7e4db13d50bedd3c0f1baa788f5949065c0d75197a4b42d387150"} Oct 04 04:48:00 crc kubenswrapper[4575]: I1004 04:48:00.342691 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerStarted","Data":"4b26e79e1839fdd1289f11fc53feb90507b8de4a80de2ada3dfce62af1de82b0"} Oct 04 04:48:00 crc kubenswrapper[4575]: I1004 04:48:00.342710 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerStarted","Data":"048d2bddce17a2e27a314a379f8a1322502047d08b1a8706c1a58403b61065e3"} Oct 04 04:48:00 crc kubenswrapper[4575]: I1004 04:48:00.342736 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerStarted","Data":"7323f7e0d3c384f172379cad66cc8b60fad81bbb49ae43a281af719d87cd5761"} Oct 04 04:48:00 crc kubenswrapper[4575]: I1004 04:48:00.342754 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerStarted","Data":"4b313f34fae5423a949a332c9107d34fc35e287abf06ce6e843ec306b2fc1585"} Oct 04 04:48:02 crc kubenswrapper[4575]: I1004 04:48:02.356487 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerStarted","Data":"4ae3a54da06797f7ca3c30cd7bab1f33a4244b23b4ca0f62c26d16b81818d8e9"} Oct 04 04:48:05 crc kubenswrapper[4575]: I1004 04:48:05.374978 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" event={"ID":"94b8f7bb-4568-4d39-a49e-7df9e84ae94d","Type":"ContainerStarted","Data":"b3918f35f37fe65ccccb21b6e0c88dbfd8116d2c0b21ed4ff4f534f9722bd053"} Oct 04 04:48:05 crc kubenswrapper[4575]: I1004 04:48:05.375328 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:48:05 crc kubenswrapper[4575]: I1004 04:48:05.375353 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:48:05 crc kubenswrapper[4575]: I1004 04:48:05.402440 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:48:05 crc kubenswrapper[4575]: I1004 04:48:05.410958 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" podStartSLOduration=7.410944507 podStartE2EDuration="7.410944507s" podCreationTimestamp="2025-10-04 04:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:48:05.410633708 +0000 UTC m=+836.739192532" watchObservedRunningTime="2025-10-04 04:48:05.410944507 +0000 UTC m=+836.739503321" Oct 04 04:48:06 crc kubenswrapper[4575]: I1004 04:48:06.381365 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:48:06 crc kubenswrapper[4575]: I1004 04:48:06.415058 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:48:08 crc kubenswrapper[4575]: I1004 04:48:08.446015 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:48:08 crc kubenswrapper[4575]: I1004 04:48:08.446091 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:48:08 crc kubenswrapper[4575]: I1004 04:48:08.446144 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:48:08 crc kubenswrapper[4575]: I1004 04:48:08.446725 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6a6d9d656e937f48c2e0c38c03877604da45131d454048196bae4e1a3e60a62"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:48:08 crc kubenswrapper[4575]: I1004 04:48:08.446790 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://e6a6d9d656e937f48c2e0c38c03877604da45131d454048196bae4e1a3e60a62" gracePeriod=600 Oct 04 04:48:09 crc kubenswrapper[4575]: I1004 04:48:09.412031 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="e6a6d9d656e937f48c2e0c38c03877604da45131d454048196bae4e1a3e60a62" exitCode=0 Oct 04 04:48:09 crc kubenswrapper[4575]: I1004 04:48:09.412107 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"e6a6d9d656e937f48c2e0c38c03877604da45131d454048196bae4e1a3e60a62"} Oct 04 04:48:09 crc kubenswrapper[4575]: I1004 04:48:09.412921 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"e6585e9613991e3a385e6fad309ba0255360c49e2f3c2664f8dc4f993639cb30"} Oct 04 04:48:09 crc kubenswrapper[4575]: I1004 04:48:09.412944 4575 scope.go:117] "RemoveContainer" containerID="a31a363eba214bd5be24aa256fc82e8758ebfb0d8fe15168b3bda5e51827dda0" Oct 04 04:48:28 crc kubenswrapper[4575]: I1004 04:48:28.435896 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jd8rt" Oct 04 04:48:40 crc kubenswrapper[4575]: I1004 04:48:40.894744 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx"] Oct 04 04:48:40 crc kubenswrapper[4575]: I1004 04:48:40.896654 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:40 crc kubenswrapper[4575]: I1004 04:48:40.899159 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 04:48:40 crc kubenswrapper[4575]: I1004 04:48:40.906248 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx"] Oct 04 04:48:40 crc kubenswrapper[4575]: I1004 04:48:40.950943 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:40 crc kubenswrapper[4575]: I1004 04:48:40.951382 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:40 crc kubenswrapper[4575]: I1004 04:48:40.951576 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl7js\" (UniqueName: \"kubernetes.io/projected/7de5d540-783e-42d5-89fd-b2a7185681b2-kube-api-access-fl7js\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:41 crc kubenswrapper[4575]: I1004 04:48:41.053997 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl7js\" (UniqueName: \"kubernetes.io/projected/7de5d540-783e-42d5-89fd-b2a7185681b2-kube-api-access-fl7js\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:41 crc kubenswrapper[4575]: I1004 04:48:41.054168 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:41 crc kubenswrapper[4575]: I1004 04:48:41.054229 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:41 crc kubenswrapper[4575]: I1004 04:48:41.055100 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:41 crc kubenswrapper[4575]: I1004 04:48:41.055247 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:41 crc kubenswrapper[4575]: I1004 04:48:41.078522 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl7js\" (UniqueName: \"kubernetes.io/projected/7de5d540-783e-42d5-89fd-b2a7185681b2-kube-api-access-fl7js\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:41 crc kubenswrapper[4575]: I1004 04:48:41.213351 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:41 crc kubenswrapper[4575]: I1004 04:48:41.607350 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx"] Oct 04 04:48:42 crc kubenswrapper[4575]: I1004 04:48:42.590506 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" event={"ID":"7de5d540-783e-42d5-89fd-b2a7185681b2","Type":"ContainerStarted","Data":"3217b000dd0ab13a4eede8287dc89e7f11ce103012680e8abbe06802680d9215"} Oct 04 04:48:42 crc kubenswrapper[4575]: I1004 04:48:42.590564 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" event={"ID":"7de5d540-783e-42d5-89fd-b2a7185681b2","Type":"ContainerStarted","Data":"3b4f7fb377de1369fe9e28465bf68023eb2cf99ca5490260f91d1408d76e333b"} Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.124641 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m528m"] Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.126354 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.156401 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m528m"] Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.190735 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlbvh\" (UniqueName: \"kubernetes.io/projected/4061ed3b-4632-41ff-b520-a7d8d56947ad-kube-api-access-xlbvh\") pod \"redhat-operators-m528m\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.190811 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-catalog-content\") pod \"redhat-operators-m528m\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.190848 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-utilities\") pod \"redhat-operators-m528m\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.292250 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlbvh\" (UniqueName: \"kubernetes.io/projected/4061ed3b-4632-41ff-b520-a7d8d56947ad-kube-api-access-xlbvh\") pod \"redhat-operators-m528m\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.292308 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-catalog-content\") pod \"redhat-operators-m528m\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.292348 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-utilities\") pod \"redhat-operators-m528m\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.292819 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-utilities\") pod \"redhat-operators-m528m\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.292922 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-catalog-content\") pod \"redhat-operators-m528m\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.321460 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlbvh\" (UniqueName: \"kubernetes.io/projected/4061ed3b-4632-41ff-b520-a7d8d56947ad-kube-api-access-xlbvh\") pod \"redhat-operators-m528m\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.441906 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.609351 4575 generic.go:334] "Generic (PLEG): container finished" podID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerID="3217b000dd0ab13a4eede8287dc89e7f11ce103012680e8abbe06802680d9215" exitCode=0 Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.609565 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" event={"ID":"7de5d540-783e-42d5-89fd-b2a7185681b2","Type":"ContainerDied","Data":"3217b000dd0ab13a4eede8287dc89e7f11ce103012680e8abbe06802680d9215"} Oct 04 04:48:43 crc kubenswrapper[4575]: I1004 04:48:43.878099 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m528m"] Oct 04 04:48:43 crc kubenswrapper[4575]: W1004 04:48:43.889937 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4061ed3b_4632_41ff_b520_a7d8d56947ad.slice/crio-57e3cdf47e797f9aeb60b3c4f90b57073fcf4d04524e7ecfe9c6c19a527c7bf4 WatchSource:0}: Error finding container 57e3cdf47e797f9aeb60b3c4f90b57073fcf4d04524e7ecfe9c6c19a527c7bf4: Status 404 returned error can't find the container with id 57e3cdf47e797f9aeb60b3c4f90b57073fcf4d04524e7ecfe9c6c19a527c7bf4 Oct 04 04:48:44 crc kubenswrapper[4575]: I1004 04:48:44.615479 4575 generic.go:334] "Generic (PLEG): container finished" podID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerID="c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def" exitCode=0 Oct 04 04:48:44 crc kubenswrapper[4575]: I1004 04:48:44.615554 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m528m" event={"ID":"4061ed3b-4632-41ff-b520-a7d8d56947ad","Type":"ContainerDied","Data":"c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def"} Oct 04 04:48:44 crc kubenswrapper[4575]: I1004 04:48:44.615880 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m528m" event={"ID":"4061ed3b-4632-41ff-b520-a7d8d56947ad","Type":"ContainerStarted","Data":"57e3cdf47e797f9aeb60b3c4f90b57073fcf4d04524e7ecfe9c6c19a527c7bf4"} Oct 04 04:48:45 crc kubenswrapper[4575]: I1004 04:48:45.621568 4575 generic.go:334] "Generic (PLEG): container finished" podID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerID="252ddca238ae1cc593a2bd500c3a4d0b5844e0120e9d851bbe4384034342382c" exitCode=0 Oct 04 04:48:45 crc kubenswrapper[4575]: I1004 04:48:45.621624 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" event={"ID":"7de5d540-783e-42d5-89fd-b2a7185681b2","Type":"ContainerDied","Data":"252ddca238ae1cc593a2bd500c3a4d0b5844e0120e9d851bbe4384034342382c"} Oct 04 04:48:46 crc kubenswrapper[4575]: I1004 04:48:46.629013 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" event={"ID":"7de5d540-783e-42d5-89fd-b2a7185681b2","Type":"ContainerStarted","Data":"fabc2c66ccfbaf109a67396e2481318d5c9c061499ba3448b30c974796c7219e"} Oct 04 04:48:46 crc kubenswrapper[4575]: I1004 04:48:46.633148 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m528m" event={"ID":"4061ed3b-4632-41ff-b520-a7d8d56947ad","Type":"ContainerStarted","Data":"468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d"} Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.639077 4575 generic.go:334] "Generic (PLEG): container finished" podID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerID="468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d" exitCode=0 Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.639151 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m528m" event={"ID":"4061ed3b-4632-41ff-b520-a7d8d56947ad","Type":"ContainerDied","Data":"468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d"} Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.645633 4575 generic.go:334] "Generic (PLEG): container finished" podID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerID="fabc2c66ccfbaf109a67396e2481318d5c9c061499ba3448b30c974796c7219e" exitCode=0 Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.645765 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" event={"ID":"7de5d540-783e-42d5-89fd-b2a7185681b2","Type":"ContainerDied","Data":"fabc2c66ccfbaf109a67396e2481318d5c9c061499ba3448b30c974796c7219e"} Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.882335 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.960546 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl7js\" (UniqueName: \"kubernetes.io/projected/7de5d540-783e-42d5-89fd-b2a7185681b2-kube-api-access-fl7js\") pod \"7de5d540-783e-42d5-89fd-b2a7185681b2\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.960744 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-bundle\") pod \"7de5d540-783e-42d5-89fd-b2a7185681b2\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.960765 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-util\") pod \"7de5d540-783e-42d5-89fd-b2a7185681b2\" (UID: \"7de5d540-783e-42d5-89fd-b2a7185681b2\") " Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.961649 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-bundle" (OuterVolumeSpecName: "bundle") pod "7de5d540-783e-42d5-89fd-b2a7185681b2" (UID: "7de5d540-783e-42d5-89fd-b2a7185681b2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:47 crc kubenswrapper[4575]: I1004 04:48:47.970644 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7de5d540-783e-42d5-89fd-b2a7185681b2-kube-api-access-fl7js" (OuterVolumeSpecName: "kube-api-access-fl7js") pod "7de5d540-783e-42d5-89fd-b2a7185681b2" (UID: "7de5d540-783e-42d5-89fd-b2a7185681b2"). InnerVolumeSpecName "kube-api-access-fl7js". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:48 crc kubenswrapper[4575]: I1004 04:48:48.062291 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl7js\" (UniqueName: \"kubernetes.io/projected/7de5d540-783e-42d5-89fd-b2a7185681b2-kube-api-access-fl7js\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:48 crc kubenswrapper[4575]: I1004 04:48:48.062359 4575 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:48 crc kubenswrapper[4575]: I1004 04:48:48.368493 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-util" (OuterVolumeSpecName: "util") pod "7de5d540-783e-42d5-89fd-b2a7185681b2" (UID: "7de5d540-783e-42d5-89fd-b2a7185681b2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:48 crc kubenswrapper[4575]: I1004 04:48:48.466271 4575 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7de5d540-783e-42d5-89fd-b2a7185681b2-util\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:48 crc kubenswrapper[4575]: I1004 04:48:48.653456 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" event={"ID":"7de5d540-783e-42d5-89fd-b2a7185681b2","Type":"ContainerDied","Data":"3b4f7fb377de1369fe9e28465bf68023eb2cf99ca5490260f91d1408d76e333b"} Oct 04 04:48:48 crc kubenswrapper[4575]: I1004 04:48:48.653501 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b4f7fb377de1369fe9e28465bf68023eb2cf99ca5490260f91d1408d76e333b" Oct 04 04:48:48 crc kubenswrapper[4575]: I1004 04:48:48.653572 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx" Oct 04 04:48:49 crc kubenswrapper[4575]: I1004 04:48:49.662799 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m528m" event={"ID":"4061ed3b-4632-41ff-b520-a7d8d56947ad","Type":"ContainerStarted","Data":"28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b"} Oct 04 04:48:49 crc kubenswrapper[4575]: I1004 04:48:49.691775 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m528m" podStartSLOduration=2.456666542 podStartE2EDuration="6.691756474s" podCreationTimestamp="2025-10-04 04:48:43 +0000 UTC" firstStartedPulling="2025-10-04 04:48:44.617359166 +0000 UTC m=+875.945917980" lastFinishedPulling="2025-10-04 04:48:48.852449098 +0000 UTC m=+880.181007912" observedRunningTime="2025-10-04 04:48:49.687971266 +0000 UTC m=+881.016530100" watchObservedRunningTime="2025-10-04 04:48:49.691756474 +0000 UTC m=+881.020315298" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.738523 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-69wc7"] Oct 04 04:48:51 crc kubenswrapper[4575]: E1004 04:48:51.739027 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerName="pull" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.739041 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerName="pull" Oct 04 04:48:51 crc kubenswrapper[4575]: E1004 04:48:51.739056 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerName="extract" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.739064 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerName="extract" Oct 04 04:48:51 crc kubenswrapper[4575]: E1004 04:48:51.739078 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerName="util" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.739084 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerName="util" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.739188 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="7de5d540-783e-42d5-89fd-b2a7185681b2" containerName="extract" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.739567 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-69wc7" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.743539 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.743891 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.744262 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-f9b7k" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.777066 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-69wc7"] Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.808716 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn748\" (UniqueName: \"kubernetes.io/projected/2a669f16-2ea1-485e-9941-4e0cbc92dff1-kube-api-access-hn748\") pod \"nmstate-operator-858ddd8f98-69wc7\" (UID: \"2a669f16-2ea1-485e-9941-4e0cbc92dff1\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-69wc7" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.910126 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn748\" (UniqueName: \"kubernetes.io/projected/2a669f16-2ea1-485e-9941-4e0cbc92dff1-kube-api-access-hn748\") pod \"nmstate-operator-858ddd8f98-69wc7\" (UID: \"2a669f16-2ea1-485e-9941-4e0cbc92dff1\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-69wc7" Oct 04 04:48:51 crc kubenswrapper[4575]: I1004 04:48:51.940093 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn748\" (UniqueName: \"kubernetes.io/projected/2a669f16-2ea1-485e-9941-4e0cbc92dff1-kube-api-access-hn748\") pod \"nmstate-operator-858ddd8f98-69wc7\" (UID: \"2a669f16-2ea1-485e-9941-4e0cbc92dff1\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-69wc7" Oct 04 04:48:52 crc kubenswrapper[4575]: I1004 04:48:52.052956 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-69wc7" Oct 04 04:48:52 crc kubenswrapper[4575]: I1004 04:48:52.502610 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-69wc7"] Oct 04 04:48:52 crc kubenswrapper[4575]: W1004 04:48:52.510325 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a669f16_2ea1_485e_9941_4e0cbc92dff1.slice/crio-c870e1880fcb7853b857f587157d7d91a32600daad33b0c4a562252c94c2dd1b WatchSource:0}: Error finding container c870e1880fcb7853b857f587157d7d91a32600daad33b0c4a562252c94c2dd1b: Status 404 returned error can't find the container with id c870e1880fcb7853b857f587157d7d91a32600daad33b0c4a562252c94c2dd1b Oct 04 04:48:52 crc kubenswrapper[4575]: I1004 04:48:52.683391 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-69wc7" event={"ID":"2a669f16-2ea1-485e-9941-4e0cbc92dff1","Type":"ContainerStarted","Data":"c870e1880fcb7853b857f587157d7d91a32600daad33b0c4a562252c94c2dd1b"} Oct 04 04:48:54 crc kubenswrapper[4575]: I1004 04:48:53.442703 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:54 crc kubenswrapper[4575]: I1004 04:48:53.442764 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:54 crc kubenswrapper[4575]: I1004 04:48:53.480696 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:54 crc kubenswrapper[4575]: I1004 04:48:53.737033 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:55 crc kubenswrapper[4575]: I1004 04:48:55.722773 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m528m"] Oct 04 04:48:55 crc kubenswrapper[4575]: I1004 04:48:55.723001 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m528m" podUID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerName="registry-server" containerID="cri-o://28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b" gracePeriod=2 Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.146239 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.175465 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlbvh\" (UniqueName: \"kubernetes.io/projected/4061ed3b-4632-41ff-b520-a7d8d56947ad-kube-api-access-xlbvh\") pod \"4061ed3b-4632-41ff-b520-a7d8d56947ad\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.175627 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-utilities\") pod \"4061ed3b-4632-41ff-b520-a7d8d56947ad\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.175675 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-catalog-content\") pod \"4061ed3b-4632-41ff-b520-a7d8d56947ad\" (UID: \"4061ed3b-4632-41ff-b520-a7d8d56947ad\") " Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.176560 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-utilities" (OuterVolumeSpecName: "utilities") pod "4061ed3b-4632-41ff-b520-a7d8d56947ad" (UID: "4061ed3b-4632-41ff-b520-a7d8d56947ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.183262 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4061ed3b-4632-41ff-b520-a7d8d56947ad-kube-api-access-xlbvh" (OuterVolumeSpecName: "kube-api-access-xlbvh") pod "4061ed3b-4632-41ff-b520-a7d8d56947ad" (UID: "4061ed3b-4632-41ff-b520-a7d8d56947ad"). InnerVolumeSpecName "kube-api-access-xlbvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.262537 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4061ed3b-4632-41ff-b520-a7d8d56947ad" (UID: "4061ed3b-4632-41ff-b520-a7d8d56947ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.277514 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.277576 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4061ed3b-4632-41ff-b520-a7d8d56947ad-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.277613 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlbvh\" (UniqueName: \"kubernetes.io/projected/4061ed3b-4632-41ff-b520-a7d8d56947ad-kube-api-access-xlbvh\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.717151 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-69wc7" event={"ID":"2a669f16-2ea1-485e-9941-4e0cbc92dff1","Type":"ContainerStarted","Data":"fe9ca71f5c5a0acbf2273c646d21cc204e3491bde39be729a7f9e37fe9d0d5f4"} Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.720001 4575 generic.go:334] "Generic (PLEG): container finished" podID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerID="28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b" exitCode=0 Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.720036 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m528m" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.720044 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m528m" event={"ID":"4061ed3b-4632-41ff-b520-a7d8d56947ad","Type":"ContainerDied","Data":"28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b"} Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.720077 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m528m" event={"ID":"4061ed3b-4632-41ff-b520-a7d8d56947ad","Type":"ContainerDied","Data":"57e3cdf47e797f9aeb60b3c4f90b57073fcf4d04524e7ecfe9c6c19a527c7bf4"} Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.720102 4575 scope.go:117] "RemoveContainer" containerID="28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.740395 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-69wc7" podStartSLOduration=2.584523442 podStartE2EDuration="5.739725172s" podCreationTimestamp="2025-10-04 04:48:51 +0000 UTC" firstStartedPulling="2025-10-04 04:48:52.51260758 +0000 UTC m=+883.841166394" lastFinishedPulling="2025-10-04 04:48:55.66780931 +0000 UTC m=+886.996368124" observedRunningTime="2025-10-04 04:48:56.733685959 +0000 UTC m=+888.062244783" watchObservedRunningTime="2025-10-04 04:48:56.739725172 +0000 UTC m=+888.068284026" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.753829 4575 scope.go:117] "RemoveContainer" containerID="468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.768801 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m528m"] Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.781263 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m528m"] Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.799880 4575 scope.go:117] "RemoveContainer" containerID="c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.816732 4575 scope.go:117] "RemoveContainer" containerID="28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b" Oct 04 04:48:56 crc kubenswrapper[4575]: E1004 04:48:56.818801 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b\": container with ID starting with 28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b not found: ID does not exist" containerID="28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.818842 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b"} err="failed to get container status \"28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b\": rpc error: code = NotFound desc = could not find container \"28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b\": container with ID starting with 28cf561c6021a8c055ff2b226ff232ab56baf90a53dd070180c58aaf9dc3ae4b not found: ID does not exist" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.818869 4575 scope.go:117] "RemoveContainer" containerID="468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d" Oct 04 04:48:56 crc kubenswrapper[4575]: E1004 04:48:56.819391 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d\": container with ID starting with 468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d not found: ID does not exist" containerID="468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.819478 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d"} err="failed to get container status \"468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d\": rpc error: code = NotFound desc = could not find container \"468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d\": container with ID starting with 468886d257a17d4c02e7e244372f22070502187d8cf9cd7602f19c38948dca5d not found: ID does not exist" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.819518 4575 scope.go:117] "RemoveContainer" containerID="c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def" Oct 04 04:48:56 crc kubenswrapper[4575]: E1004 04:48:56.820090 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def\": container with ID starting with c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def not found: ID does not exist" containerID="c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def" Oct 04 04:48:56 crc kubenswrapper[4575]: I1004 04:48:56.820118 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def"} err="failed to get container status \"c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def\": rpc error: code = NotFound desc = could not find container \"c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def\": container with ID starting with c37388c1e4cfa189e9176d81bd9ebbef0010e65b19055d134841849064746def not found: ID does not exist" Oct 04 04:48:57 crc kubenswrapper[4575]: I1004 04:48:57.318871 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4061ed3b-4632-41ff-b520-a7d8d56947ad" path="/var/lib/kubelet/pods/4061ed3b-4632-41ff-b520-a7d8d56947ad/volumes" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.662738 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv"] Oct 04 04:49:01 crc kubenswrapper[4575]: E1004 04:49:01.663551 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerName="registry-server" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.663569 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerName="registry-server" Oct 04 04:49:01 crc kubenswrapper[4575]: E1004 04:49:01.663598 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerName="extract-utilities" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.663605 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerName="extract-utilities" Oct 04 04:49:01 crc kubenswrapper[4575]: E1004 04:49:01.663617 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerName="extract-content" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.663624 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerName="extract-content" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.663771 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="4061ed3b-4632-41ff-b520-a7d8d56947ad" containerName="registry-server" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.664613 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.666727 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-jwh2b" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.684548 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv"] Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.689255 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5"] Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.690271 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.698085 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.724062 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5"] Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.763716 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-jhfj5\" (UID: \"fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.763805 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz5dl\" (UniqueName: \"kubernetes.io/projected/0155bc5a-f5fc-4e45-b557-d950e8d95ff0-kube-api-access-pz5dl\") pod \"nmstate-metrics-fdff9cb8d-5d2mv\" (UID: \"0155bc5a-f5fc-4e45-b557-d950e8d95ff0\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.763835 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8df4\" (UniqueName: \"kubernetes.io/projected/fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d-kube-api-access-j8df4\") pod \"nmstate-webhook-6cdbc54649-jhfj5\" (UID: \"fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.781786 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-d4cmv"] Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.788722 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.865187 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/38de0af5-93f8-4b73-aadd-9c2932e4de86-ovs-socket\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.865263 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgbr6\" (UniqueName: \"kubernetes.io/projected/38de0af5-93f8-4b73-aadd-9c2932e4de86-kube-api-access-qgbr6\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.865292 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/38de0af5-93f8-4b73-aadd-9c2932e4de86-nmstate-lock\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.865328 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-jhfj5\" (UID: \"fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:01 crc kubenswrapper[4575]: E1004 04:49:01.865469 4575 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 04 04:49:01 crc kubenswrapper[4575]: E1004 04:49:01.865544 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d-tls-key-pair podName:fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d nodeName:}" failed. No retries permitted until 2025-10-04 04:49:02.365520237 +0000 UTC m=+893.694079221 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d-tls-key-pair") pod "nmstate-webhook-6cdbc54649-jhfj5" (UID: "fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d") : secret "openshift-nmstate-webhook" not found Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.865708 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz5dl\" (UniqueName: \"kubernetes.io/projected/0155bc5a-f5fc-4e45-b557-d950e8d95ff0-kube-api-access-pz5dl\") pod \"nmstate-metrics-fdff9cb8d-5d2mv\" (UID: \"0155bc5a-f5fc-4e45-b557-d950e8d95ff0\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.865816 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8df4\" (UniqueName: \"kubernetes.io/projected/fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d-kube-api-access-j8df4\") pod \"nmstate-webhook-6cdbc54649-jhfj5\" (UID: \"fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.865871 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/38de0af5-93f8-4b73-aadd-9c2932e4de86-dbus-socket\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.898040 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8df4\" (UniqueName: \"kubernetes.io/projected/fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d-kube-api-access-j8df4\") pod \"nmstate-webhook-6cdbc54649-jhfj5\" (UID: \"fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.907906 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz5dl\" (UniqueName: \"kubernetes.io/projected/0155bc5a-f5fc-4e45-b557-d950e8d95ff0-kube-api-access-pz5dl\") pod \"nmstate-metrics-fdff9cb8d-5d2mv\" (UID: \"0155bc5a-f5fc-4e45-b557-d950e8d95ff0\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.935158 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx"] Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.936178 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.940021 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.943494 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.943773 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-zd6db" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.943042 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx"] Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.967770 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/38de0af5-93f8-4b73-aadd-9c2932e4de86-dbus-socket\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.967869 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/38de0af5-93f8-4b73-aadd-9c2932e4de86-ovs-socket\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.967905 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgbr6\" (UniqueName: \"kubernetes.io/projected/38de0af5-93f8-4b73-aadd-9c2932e4de86-kube-api-access-qgbr6\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.967922 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/38de0af5-93f8-4b73-aadd-9c2932e4de86-nmstate-lock\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.968026 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/38de0af5-93f8-4b73-aadd-9c2932e4de86-nmstate-lock\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.968477 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/38de0af5-93f8-4b73-aadd-9c2932e4de86-dbus-socket\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.968499 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/38de0af5-93f8-4b73-aadd-9c2932e4de86-ovs-socket\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.992530 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgbr6\" (UniqueName: \"kubernetes.io/projected/38de0af5-93f8-4b73-aadd-9c2932e4de86-kube-api-access-qgbr6\") pod \"nmstate-handler-d4cmv\" (UID: \"38de0af5-93f8-4b73-aadd-9c2932e4de86\") " pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:01 crc kubenswrapper[4575]: I1004 04:49:01.999177 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.069927 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/255da756-dc3e-4ea1-946b-82593d8be138-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.069977 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4njqp\" (UniqueName: \"kubernetes.io/projected/255da756-dc3e-4ea1-946b-82593d8be138-kube-api-access-4njqp\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.070045 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/255da756-dc3e-4ea1-946b-82593d8be138-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.120920 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.171851 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/255da756-dc3e-4ea1-946b-82593d8be138-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.172315 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/255da756-dc3e-4ea1-946b-82593d8be138-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.172340 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4njqp\" (UniqueName: \"kubernetes.io/projected/255da756-dc3e-4ea1-946b-82593d8be138-kube-api-access-4njqp\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.173856 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/255da756-dc3e-4ea1-946b-82593d8be138-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: E1004 04:49:02.173961 4575 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 04 04:49:02 crc kubenswrapper[4575]: E1004 04:49:02.174059 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/255da756-dc3e-4ea1-946b-82593d8be138-plugin-serving-cert podName:255da756-dc3e-4ea1-946b-82593d8be138 nodeName:}" failed. No retries permitted until 2025-10-04 04:49:02.674037341 +0000 UTC m=+894.002596155 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/255da756-dc3e-4ea1-946b-82593d8be138-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-6tqcx" (UID: "255da756-dc3e-4ea1-946b-82593d8be138") : secret "plugin-serving-cert" not found Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.214303 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7bfc466b86-rs5lk"] Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.221149 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.227522 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4njqp\" (UniqueName: \"kubernetes.io/projected/255da756-dc3e-4ea1-946b-82593d8be138-kube-api-access-4njqp\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.260443 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7bfc466b86-rs5lk"] Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.274471 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cac2eb70-84d2-4b14-a346-091eb45a5d90-console-serving-cert\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.274511 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-console-config\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.274566 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-trusted-ca-bundle\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.274657 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cac2eb70-84d2-4b14-a346-091eb45a5d90-console-oauth-config\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.274797 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-service-ca\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.274904 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-oauth-serving-cert\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.274929 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pks4f\" (UniqueName: \"kubernetes.io/projected/cac2eb70-84d2-4b14-a346-091eb45a5d90-kube-api-access-pks4f\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.376504 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cac2eb70-84d2-4b14-a346-091eb45a5d90-console-serving-cert\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.376560 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-console-config\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.376615 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-trusted-ca-bundle\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.376702 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-jhfj5\" (UID: \"fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.377902 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-console-config\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.378071 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv"] Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.380014 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-jhfj5\" (UID: \"fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.380211 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cac2eb70-84d2-4b14-a346-091eb45a5d90-console-serving-cert\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.381027 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cac2eb70-84d2-4b14-a346-091eb45a5d90-console-oauth-config\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.390061 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-trusted-ca-bundle\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.376727 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cac2eb70-84d2-4b14-a346-091eb45a5d90-console-oauth-config\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.390239 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-service-ca\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.390290 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-oauth-serving-cert\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.390310 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pks4f\" (UniqueName: \"kubernetes.io/projected/cac2eb70-84d2-4b14-a346-091eb45a5d90-kube-api-access-pks4f\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.391528 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-service-ca\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.392228 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cac2eb70-84d2-4b14-a346-091eb45a5d90-oauth-serving-cert\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.439193 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pks4f\" (UniqueName: \"kubernetes.io/projected/cac2eb70-84d2-4b14-a346-091eb45a5d90-kube-api-access-pks4f\") pod \"console-7bfc466b86-rs5lk\" (UID: \"cac2eb70-84d2-4b14-a346-091eb45a5d90\") " pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.584219 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.616092 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.694553 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/255da756-dc3e-4ea1-946b-82593d8be138-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.698814 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/255da756-dc3e-4ea1-946b-82593d8be138-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-6tqcx\" (UID: \"255da756-dc3e-4ea1-946b-82593d8be138\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.780994 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv" event={"ID":"0155bc5a-f5fc-4e45-b557-d950e8d95ff0","Type":"ContainerStarted","Data":"458240333bcbcd5cc5056bf94845100449208bebe7115adb0846a625df1688a7"} Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.782017 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-d4cmv" event={"ID":"38de0af5-93f8-4b73-aadd-9c2932e4de86","Type":"ContainerStarted","Data":"81358cbbb24e5c0c54d96fe310a52785b855a6dc46388a54a0a5c29455137f02"} Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.813671 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7bfc466b86-rs5lk"] Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.880656 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5"] Oct 04 04:49:02 crc kubenswrapper[4575]: I1004 04:49:02.883920 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" Oct 04 04:49:02 crc kubenswrapper[4575]: W1004 04:49:02.890522 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc5b01a0_9da5_4b69_b0c0_43f01cf3ab1d.slice/crio-be61db588dd4de543a87a3cdf78ae56a7c300d3a176052a1cca40fcdd3c90650 WatchSource:0}: Error finding container be61db588dd4de543a87a3cdf78ae56a7c300d3a176052a1cca40fcdd3c90650: Status 404 returned error can't find the container with id be61db588dd4de543a87a3cdf78ae56a7c300d3a176052a1cca40fcdd3c90650 Oct 04 04:49:03 crc kubenswrapper[4575]: I1004 04:49:03.157490 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx"] Oct 04 04:49:03 crc kubenswrapper[4575]: I1004 04:49:03.788714 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" event={"ID":"255da756-dc3e-4ea1-946b-82593d8be138","Type":"ContainerStarted","Data":"ae2677051f9ed8ec940bd6a40ebebcb4fc22a469715abe0b486f927b92eb9fd1"} Oct 04 04:49:03 crc kubenswrapper[4575]: I1004 04:49:03.790420 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" event={"ID":"fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d","Type":"ContainerStarted","Data":"be61db588dd4de543a87a3cdf78ae56a7c300d3a176052a1cca40fcdd3c90650"} Oct 04 04:49:03 crc kubenswrapper[4575]: I1004 04:49:03.791837 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7bfc466b86-rs5lk" event={"ID":"cac2eb70-84d2-4b14-a346-091eb45a5d90","Type":"ContainerStarted","Data":"c9323d134e0c5748edb8c3cfcd7ad7081784d2acb5869d0887b323ec7eb2d58b"} Oct 04 04:49:03 crc kubenswrapper[4575]: I1004 04:49:03.792049 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7bfc466b86-rs5lk" event={"ID":"cac2eb70-84d2-4b14-a346-091eb45a5d90","Type":"ContainerStarted","Data":"19ec72b2205342d976952089c64c22f1f0fddf9dc4e8b1049259ad0e427b11ac"} Oct 04 04:49:03 crc kubenswrapper[4575]: I1004 04:49:03.813479 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7bfc466b86-rs5lk" podStartSLOduration=1.81345788 podStartE2EDuration="1.81345788s" podCreationTimestamp="2025-10-04 04:49:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:49:03.809993121 +0000 UTC m=+895.138551935" watchObservedRunningTime="2025-10-04 04:49:03.81345788 +0000 UTC m=+895.142016704" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.521793 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nxms9"] Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.523452 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.534468 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nxms9"] Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.654650 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tqx4\" (UniqueName: \"kubernetes.io/projected/d1d0921b-3c88-4f60-8aa1-051700e7605d-kube-api-access-4tqx4\") pod \"community-operators-nxms9\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.654731 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-catalog-content\") pod \"community-operators-nxms9\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.654827 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-utilities\") pod \"community-operators-nxms9\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.756286 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-catalog-content\") pod \"community-operators-nxms9\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.756358 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-utilities\") pod \"community-operators-nxms9\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.756398 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tqx4\" (UniqueName: \"kubernetes.io/projected/d1d0921b-3c88-4f60-8aa1-051700e7605d-kube-api-access-4tqx4\") pod \"community-operators-nxms9\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.756837 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-utilities\") pod \"community-operators-nxms9\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.756837 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-catalog-content\") pod \"community-operators-nxms9\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.775290 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tqx4\" (UniqueName: \"kubernetes.io/projected/d1d0921b-3c88-4f60-8aa1-051700e7605d-kube-api-access-4tqx4\") pod \"community-operators-nxms9\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:06 crc kubenswrapper[4575]: I1004 04:49:06.849806 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:07 crc kubenswrapper[4575]: I1004 04:49:07.127687 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nxms9"] Oct 04 04:49:07 crc kubenswrapper[4575]: W1004 04:49:07.137221 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1d0921b_3c88_4f60_8aa1_051700e7605d.slice/crio-8fadb49c81cfbe31e648df47baaa32fab1ff5519429a58b12df5094f795d204e WatchSource:0}: Error finding container 8fadb49c81cfbe31e648df47baaa32fab1ff5519429a58b12df5094f795d204e: Status 404 returned error can't find the container with id 8fadb49c81cfbe31e648df47baaa32fab1ff5519429a58b12df5094f795d204e Oct 04 04:49:07 crc kubenswrapper[4575]: I1004 04:49:07.815325 4575 generic.go:334] "Generic (PLEG): container finished" podID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerID="908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5" exitCode=0 Oct 04 04:49:07 crc kubenswrapper[4575]: I1004 04:49:07.815442 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxms9" event={"ID":"d1d0921b-3c88-4f60-8aa1-051700e7605d","Type":"ContainerDied","Data":"908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5"} Oct 04 04:49:07 crc kubenswrapper[4575]: I1004 04:49:07.815686 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxms9" event={"ID":"d1d0921b-3c88-4f60-8aa1-051700e7605d","Type":"ContainerStarted","Data":"8fadb49c81cfbe31e648df47baaa32fab1ff5519429a58b12df5094f795d204e"} Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.321434 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6j6nz"] Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.323069 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.338330 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6j6nz"] Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.375565 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-utilities\") pod \"redhat-marketplace-6j6nz\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.375714 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-catalog-content\") pod \"redhat-marketplace-6j6nz\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.375749 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2h65\" (UniqueName: \"kubernetes.io/projected/8a880646-7084-4eb9-9341-c50e28a34ec6-kube-api-access-l2h65\") pod \"redhat-marketplace-6j6nz\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.477296 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-utilities\") pod \"redhat-marketplace-6j6nz\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.477418 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-catalog-content\") pod \"redhat-marketplace-6j6nz\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.477453 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2h65\" (UniqueName: \"kubernetes.io/projected/8a880646-7084-4eb9-9341-c50e28a34ec6-kube-api-access-l2h65\") pod \"redhat-marketplace-6j6nz\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.478022 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-utilities\") pod \"redhat-marketplace-6j6nz\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.478046 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-catalog-content\") pod \"redhat-marketplace-6j6nz\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.502776 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2h65\" (UniqueName: \"kubernetes.io/projected/8a880646-7084-4eb9-9341-c50e28a34ec6-kube-api-access-l2h65\") pod \"redhat-marketplace-6j6nz\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.642766 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.914964 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rxztc"] Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.916047 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.931046 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rxztc"] Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.984098 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-utilities\") pod \"certified-operators-rxztc\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.984150 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgfg6\" (UniqueName: \"kubernetes.io/projected/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-kube-api-access-fgfg6\") pod \"certified-operators-rxztc\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:08 crc kubenswrapper[4575]: I1004 04:49:08.984182 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-catalog-content\") pod \"certified-operators-rxztc\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.085484 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-utilities\") pod \"certified-operators-rxztc\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.085544 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgfg6\" (UniqueName: \"kubernetes.io/projected/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-kube-api-access-fgfg6\") pod \"certified-operators-rxztc\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.085571 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-catalog-content\") pod \"certified-operators-rxztc\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.086084 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-utilities\") pod \"certified-operators-rxztc\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.086095 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-catalog-content\") pod \"certified-operators-rxztc\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.111783 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgfg6\" (UniqueName: \"kubernetes.io/projected/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-kube-api-access-fgfg6\") pod \"certified-operators-rxztc\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.232631 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.586722 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6j6nz"] Oct 04 04:49:09 crc kubenswrapper[4575]: W1004 04:49:09.675466 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a880646_7084_4eb9_9341_c50e28a34ec6.slice/crio-322be36d62903f32f0220621c131791d26cd4fb46aae7e883c22826f8982e75c WatchSource:0}: Error finding container 322be36d62903f32f0220621c131791d26cd4fb46aae7e883c22826f8982e75c: Status 404 returned error can't find the container with id 322be36d62903f32f0220621c131791d26cd4fb46aae7e883c22826f8982e75c Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.833397 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6j6nz" event={"ID":"8a880646-7084-4eb9-9341-c50e28a34ec6","Type":"ContainerStarted","Data":"322be36d62903f32f0220621c131791d26cd4fb46aae7e883c22826f8982e75c"} Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.834788 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv" event={"ID":"0155bc5a-f5fc-4e45-b557-d950e8d95ff0","Type":"ContainerStarted","Data":"3b2e480c99bec98b37369afc9134ed498fc72f88357c6a537964c7e6aa764eab"} Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.835799 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" event={"ID":"255da756-dc3e-4ea1-946b-82593d8be138","Type":"ContainerStarted","Data":"326c376d49c75e3b31e4c7ed6c19cfa49c078d6530193f58aba4e19804c1162a"} Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.838144 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" event={"ID":"fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d","Type":"ContainerStarted","Data":"dbed49815bdfedf0a982d88de567e2cceaf22c2ddbeb1ffdf747951b28d26e1f"} Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.838846 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.840994 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-d4cmv" event={"ID":"38de0af5-93f8-4b73-aadd-9c2932e4de86","Type":"ContainerStarted","Data":"f9b15249bdc97603424bbfa2063de34da9087d38175919eb625636112fcbbbb9"} Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.841285 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.855642 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-6tqcx" podStartSLOduration=2.967111011 podStartE2EDuration="8.855623872s" podCreationTimestamp="2025-10-04 04:49:01 +0000 UTC" firstStartedPulling="2025-10-04 04:49:03.174954377 +0000 UTC m=+894.503513191" lastFinishedPulling="2025-10-04 04:49:09.063467238 +0000 UTC m=+900.392026052" observedRunningTime="2025-10-04 04:49:09.854384256 +0000 UTC m=+901.182943090" watchObservedRunningTime="2025-10-04 04:49:09.855623872 +0000 UTC m=+901.184182696" Oct 04 04:49:09 crc kubenswrapper[4575]: I1004 04:49:09.900651 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" podStartSLOduration=2.778894629 podStartE2EDuration="8.900629823s" podCreationTimestamp="2025-10-04 04:49:01 +0000 UTC" firstStartedPulling="2025-10-04 04:49:02.928984568 +0000 UTC m=+894.257543382" lastFinishedPulling="2025-10-04 04:49:09.050719762 +0000 UTC m=+900.379278576" observedRunningTime="2025-10-04 04:49:09.881646728 +0000 UTC m=+901.210205542" watchObservedRunningTime="2025-10-04 04:49:09.900629823 +0000 UTC m=+901.229188637" Oct 04 04:49:10 crc kubenswrapper[4575]: I1004 04:49:10.184438 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-d4cmv" podStartSLOduration=2.364394943 podStartE2EDuration="9.184410257s" podCreationTimestamp="2025-10-04 04:49:01 +0000 UTC" firstStartedPulling="2025-10-04 04:49:02.230709028 +0000 UTC m=+893.559267832" lastFinishedPulling="2025-10-04 04:49:09.050724332 +0000 UTC m=+900.379283146" observedRunningTime="2025-10-04 04:49:09.901241381 +0000 UTC m=+901.229800185" watchObservedRunningTime="2025-10-04 04:49:10.184410257 +0000 UTC m=+901.512969071" Oct 04 04:49:10 crc kubenswrapper[4575]: I1004 04:49:10.189512 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rxztc"] Oct 04 04:49:10 crc kubenswrapper[4575]: I1004 04:49:10.852921 4575 generic.go:334] "Generic (PLEG): container finished" podID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerID="4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a" exitCode=0 Oct 04 04:49:10 crc kubenswrapper[4575]: I1004 04:49:10.853037 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxms9" event={"ID":"d1d0921b-3c88-4f60-8aa1-051700e7605d","Type":"ContainerDied","Data":"4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a"} Oct 04 04:49:10 crc kubenswrapper[4575]: I1004 04:49:10.855897 4575 generic.go:334] "Generic (PLEG): container finished" podID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerID="c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626" exitCode=0 Oct 04 04:49:10 crc kubenswrapper[4575]: I1004 04:49:10.855946 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6j6nz" event={"ID":"8a880646-7084-4eb9-9341-c50e28a34ec6","Type":"ContainerDied","Data":"c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626"} Oct 04 04:49:10 crc kubenswrapper[4575]: I1004 04:49:10.866424 4575 generic.go:334] "Generic (PLEG): container finished" podID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerID="fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7" exitCode=0 Oct 04 04:49:10 crc kubenswrapper[4575]: I1004 04:49:10.868011 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxztc" event={"ID":"27c6d7db-7681-4d05-adde-48e6fb2d4eb9","Type":"ContainerDied","Data":"fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7"} Oct 04 04:49:10 crc kubenswrapper[4575]: I1004 04:49:10.868049 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxztc" event={"ID":"27c6d7db-7681-4d05-adde-48e6fb2d4eb9","Type":"ContainerStarted","Data":"2b48d93a11069007e24b62d3584d969335ee979b31d647cc1b5aac0f5b13f169"} Oct 04 04:49:12 crc kubenswrapper[4575]: I1004 04:49:12.584877 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:12 crc kubenswrapper[4575]: I1004 04:49:12.585369 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:12 crc kubenswrapper[4575]: I1004 04:49:12.590976 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:12 crc kubenswrapper[4575]: I1004 04:49:12.887561 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7bfc466b86-rs5lk" Oct 04 04:49:12 crc kubenswrapper[4575]: I1004 04:49:12.963579 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vgclk"] Oct 04 04:49:13 crc kubenswrapper[4575]: I1004 04:49:13.893565 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv" event={"ID":"0155bc5a-f5fc-4e45-b557-d950e8d95ff0","Type":"ContainerStarted","Data":"577da7556687d4423e6227102c089a29de846cba536c3bd77ad3e686df60b5dc"} Oct 04 04:49:13 crc kubenswrapper[4575]: I1004 04:49:13.895775 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxztc" event={"ID":"27c6d7db-7681-4d05-adde-48e6fb2d4eb9","Type":"ContainerStarted","Data":"6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58"} Oct 04 04:49:13 crc kubenswrapper[4575]: I1004 04:49:13.898142 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxms9" event={"ID":"d1d0921b-3c88-4f60-8aa1-051700e7605d","Type":"ContainerStarted","Data":"5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1"} Oct 04 04:49:13 crc kubenswrapper[4575]: I1004 04:49:13.900947 4575 generic.go:334] "Generic (PLEG): container finished" podID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerID="6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9" exitCode=0 Oct 04 04:49:13 crc kubenswrapper[4575]: I1004 04:49:13.901988 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6j6nz" event={"ID":"8a880646-7084-4eb9-9341-c50e28a34ec6","Type":"ContainerDied","Data":"6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9"} Oct 04 04:49:13 crc kubenswrapper[4575]: I1004 04:49:13.923497 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-5d2mv" podStartSLOduration=2.082189565 podStartE2EDuration="12.923480694s" podCreationTimestamp="2025-10-04 04:49:01 +0000 UTC" firstStartedPulling="2025-10-04 04:49:02.396953789 +0000 UTC m=+893.725512603" lastFinishedPulling="2025-10-04 04:49:13.238244918 +0000 UTC m=+904.566803732" observedRunningTime="2025-10-04 04:49:13.920449567 +0000 UTC m=+905.249008381" watchObservedRunningTime="2025-10-04 04:49:13.923480694 +0000 UTC m=+905.252039508" Oct 04 04:49:13 crc kubenswrapper[4575]: I1004 04:49:13.986963 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nxms9" podStartSLOduration=3.208012956 podStartE2EDuration="7.986947505s" podCreationTimestamp="2025-10-04 04:49:06 +0000 UTC" firstStartedPulling="2025-10-04 04:49:07.928000451 +0000 UTC m=+899.256559265" lastFinishedPulling="2025-10-04 04:49:12.706935 +0000 UTC m=+904.035493814" observedRunningTime="2025-10-04 04:49:13.986320257 +0000 UTC m=+905.314879071" watchObservedRunningTime="2025-10-04 04:49:13.986947505 +0000 UTC m=+905.315506319" Oct 04 04:49:14 crc kubenswrapper[4575]: I1004 04:49:14.908212 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6j6nz" event={"ID":"8a880646-7084-4eb9-9341-c50e28a34ec6","Type":"ContainerStarted","Data":"c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b"} Oct 04 04:49:14 crc kubenswrapper[4575]: I1004 04:49:14.909738 4575 generic.go:334] "Generic (PLEG): container finished" podID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerID="6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58" exitCode=0 Oct 04 04:49:14 crc kubenswrapper[4575]: I1004 04:49:14.909822 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxztc" event={"ID":"27c6d7db-7681-4d05-adde-48e6fb2d4eb9","Type":"ContainerDied","Data":"6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58"} Oct 04 04:49:16 crc kubenswrapper[4575]: I1004 04:49:16.850663 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:16 crc kubenswrapper[4575]: I1004 04:49:16.851184 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:16 crc kubenswrapper[4575]: I1004 04:49:16.906323 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:16 crc kubenswrapper[4575]: I1004 04:49:16.923884 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxztc" event={"ID":"27c6d7db-7681-4d05-adde-48e6fb2d4eb9","Type":"ContainerStarted","Data":"507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177"} Oct 04 04:49:16 crc kubenswrapper[4575]: I1004 04:49:16.924472 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6j6nz" podStartSLOduration=5.104017937 podStartE2EDuration="8.924454817s" podCreationTimestamp="2025-10-04 04:49:08 +0000 UTC" firstStartedPulling="2025-10-04 04:49:10.857930747 +0000 UTC m=+902.186489561" lastFinishedPulling="2025-10-04 04:49:14.678367627 +0000 UTC m=+906.006926441" observedRunningTime="2025-10-04 04:49:15.939447858 +0000 UTC m=+907.268006662" watchObservedRunningTime="2025-10-04 04:49:16.924454817 +0000 UTC m=+908.253013621" Oct 04 04:49:16 crc kubenswrapper[4575]: I1004 04:49:16.944389 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rxztc" podStartSLOduration=3.584236591 podStartE2EDuration="8.944367839s" podCreationTimestamp="2025-10-04 04:49:08 +0000 UTC" firstStartedPulling="2025-10-04 04:49:10.869160909 +0000 UTC m=+902.197719723" lastFinishedPulling="2025-10-04 04:49:16.229292157 +0000 UTC m=+907.557850971" observedRunningTime="2025-10-04 04:49:16.942854495 +0000 UTC m=+908.271413329" watchObservedRunningTime="2025-10-04 04:49:16.944367839 +0000 UTC m=+908.272926663" Oct 04 04:49:17 crc kubenswrapper[4575]: I1004 04:49:17.144196 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-d4cmv" Oct 04 04:49:18 crc kubenswrapper[4575]: I1004 04:49:18.643645 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:18 crc kubenswrapper[4575]: I1004 04:49:18.645299 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:18 crc kubenswrapper[4575]: I1004 04:49:18.692555 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:19 crc kubenswrapper[4575]: I1004 04:49:19.232783 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:19 crc kubenswrapper[4575]: I1004 04:49:19.232830 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:19 crc kubenswrapper[4575]: I1004 04:49:19.273311 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:22 crc kubenswrapper[4575]: I1004 04:49:22.622500 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-jhfj5" Oct 04 04:49:26 crc kubenswrapper[4575]: I1004 04:49:26.900650 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:26 crc kubenswrapper[4575]: I1004 04:49:26.953770 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nxms9"] Oct 04 04:49:26 crc kubenswrapper[4575]: I1004 04:49:26.976779 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nxms9" podUID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerName="registry-server" containerID="cri-o://5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1" gracePeriod=2 Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.330196 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.460768 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-catalog-content\") pod \"d1d0921b-3c88-4f60-8aa1-051700e7605d\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.461273 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tqx4\" (UniqueName: \"kubernetes.io/projected/d1d0921b-3c88-4f60-8aa1-051700e7605d-kube-api-access-4tqx4\") pod \"d1d0921b-3c88-4f60-8aa1-051700e7605d\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.461341 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-utilities\") pod \"d1d0921b-3c88-4f60-8aa1-051700e7605d\" (UID: \"d1d0921b-3c88-4f60-8aa1-051700e7605d\") " Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.462101 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-utilities" (OuterVolumeSpecName: "utilities") pod "d1d0921b-3c88-4f60-8aa1-051700e7605d" (UID: "d1d0921b-3c88-4f60-8aa1-051700e7605d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.466405 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d0921b-3c88-4f60-8aa1-051700e7605d-kube-api-access-4tqx4" (OuterVolumeSpecName: "kube-api-access-4tqx4") pod "d1d0921b-3c88-4f60-8aa1-051700e7605d" (UID: "d1d0921b-3c88-4f60-8aa1-051700e7605d"). InnerVolumeSpecName "kube-api-access-4tqx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.504890 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1d0921b-3c88-4f60-8aa1-051700e7605d" (UID: "d1d0921b-3c88-4f60-8aa1-051700e7605d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.563194 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.563234 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tqx4\" (UniqueName: \"kubernetes.io/projected/d1d0921b-3c88-4f60-8aa1-051700e7605d-kube-api-access-4tqx4\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.563254 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d0921b-3c88-4f60-8aa1-051700e7605d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.984760 4575 generic.go:334] "Generic (PLEG): container finished" podID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerID="5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1" exitCode=0 Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.984816 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxms9" event={"ID":"d1d0921b-3c88-4f60-8aa1-051700e7605d","Type":"ContainerDied","Data":"5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1"} Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.984853 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxms9" event={"ID":"d1d0921b-3c88-4f60-8aa1-051700e7605d","Type":"ContainerDied","Data":"8fadb49c81cfbe31e648df47baaa32fab1ff5519429a58b12df5094f795d204e"} Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.984875 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxms9" Oct 04 04:49:27 crc kubenswrapper[4575]: I1004 04:49:27.984896 4575 scope.go:117] "RemoveContainer" containerID="5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1" Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.015190 4575 scope.go:117] "RemoveContainer" containerID="4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a" Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.019577 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nxms9"] Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.024988 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nxms9"] Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.039887 4575 scope.go:117] "RemoveContainer" containerID="908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5" Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.063450 4575 scope.go:117] "RemoveContainer" containerID="5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1" Oct 04 04:49:28 crc kubenswrapper[4575]: E1004 04:49:28.064650 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1\": container with ID starting with 5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1 not found: ID does not exist" containerID="5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1" Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.064712 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1"} err="failed to get container status \"5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1\": rpc error: code = NotFound desc = could not find container \"5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1\": container with ID starting with 5370150f5955c787a65f57369e53db8d7dcf88dc15756ecb714b322fe05ef3a1 not found: ID does not exist" Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.064747 4575 scope.go:117] "RemoveContainer" containerID="4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a" Oct 04 04:49:28 crc kubenswrapper[4575]: E1004 04:49:28.065220 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a\": container with ID starting with 4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a not found: ID does not exist" containerID="4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a" Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.065368 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a"} err="failed to get container status \"4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a\": rpc error: code = NotFound desc = could not find container \"4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a\": container with ID starting with 4ec6bc7dc0f98694048bd250a2028e7e350f6eada6d56e09583a057bdcf7cb5a not found: ID does not exist" Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.065483 4575 scope.go:117] "RemoveContainer" containerID="908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5" Oct 04 04:49:28 crc kubenswrapper[4575]: E1004 04:49:28.066231 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5\": container with ID starting with 908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5 not found: ID does not exist" containerID="908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5" Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.066294 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5"} err="failed to get container status \"908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5\": rpc error: code = NotFound desc = could not find container \"908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5\": container with ID starting with 908eda4769b4b83e51c337988fd9f82b05763a915927cf70b2d2d16de65a06d5 not found: ID does not exist" Oct 04 04:49:28 crc kubenswrapper[4575]: I1004 04:49:28.683500 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:29 crc kubenswrapper[4575]: I1004 04:49:29.275378 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:29 crc kubenswrapper[4575]: I1004 04:49:29.318167 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d0921b-3c88-4f60-8aa1-051700e7605d" path="/var/lib/kubelet/pods/d1d0921b-3c88-4f60-8aa1-051700e7605d/volumes" Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.132736 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6j6nz"] Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.133513 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6j6nz" podUID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerName="registry-server" containerID="cri-o://c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b" gracePeriod=2 Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.524989 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.629893 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-utilities\") pod \"8a880646-7084-4eb9-9341-c50e28a34ec6\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.629984 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2h65\" (UniqueName: \"kubernetes.io/projected/8a880646-7084-4eb9-9341-c50e28a34ec6-kube-api-access-l2h65\") pod \"8a880646-7084-4eb9-9341-c50e28a34ec6\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.630117 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-catalog-content\") pod \"8a880646-7084-4eb9-9341-c50e28a34ec6\" (UID: \"8a880646-7084-4eb9-9341-c50e28a34ec6\") " Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.631024 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-utilities" (OuterVolumeSpecName: "utilities") pod "8a880646-7084-4eb9-9341-c50e28a34ec6" (UID: "8a880646-7084-4eb9-9341-c50e28a34ec6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.636322 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a880646-7084-4eb9-9341-c50e28a34ec6-kube-api-access-l2h65" (OuterVolumeSpecName: "kube-api-access-l2h65") pod "8a880646-7084-4eb9-9341-c50e28a34ec6" (UID: "8a880646-7084-4eb9-9341-c50e28a34ec6"). InnerVolumeSpecName "kube-api-access-l2h65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.644577 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a880646-7084-4eb9-9341-c50e28a34ec6" (UID: "8a880646-7084-4eb9-9341-c50e28a34ec6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.731332 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.731364 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2h65\" (UniqueName: \"kubernetes.io/projected/8a880646-7084-4eb9-9341-c50e28a34ec6-kube-api-access-l2h65\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:31 crc kubenswrapper[4575]: I1004 04:49:31.731377 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a880646-7084-4eb9-9341-c50e28a34ec6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.009935 4575 generic.go:334] "Generic (PLEG): container finished" podID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerID="c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b" exitCode=0 Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.010309 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6j6nz" event={"ID":"8a880646-7084-4eb9-9341-c50e28a34ec6","Type":"ContainerDied","Data":"c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b"} Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.010365 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6j6nz" event={"ID":"8a880646-7084-4eb9-9341-c50e28a34ec6","Type":"ContainerDied","Data":"322be36d62903f32f0220621c131791d26cd4fb46aae7e883c22826f8982e75c"} Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.010394 4575 scope.go:117] "RemoveContainer" containerID="c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.010726 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6j6nz" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.034770 4575 scope.go:117] "RemoveContainer" containerID="6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.054810 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6j6nz"] Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.057190 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6j6nz"] Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.059951 4575 scope.go:117] "RemoveContainer" containerID="c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.079796 4575 scope.go:117] "RemoveContainer" containerID="c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b" Oct 04 04:49:32 crc kubenswrapper[4575]: E1004 04:49:32.080424 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b\": container with ID starting with c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b not found: ID does not exist" containerID="c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.080462 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b"} err="failed to get container status \"c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b\": rpc error: code = NotFound desc = could not find container \"c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b\": container with ID starting with c40fccf83573dc17569ae4d8b77b0f198bcf8d6092a2db2da0421e7c9cd9719b not found: ID does not exist" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.080486 4575 scope.go:117] "RemoveContainer" containerID="6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9" Oct 04 04:49:32 crc kubenswrapper[4575]: E1004 04:49:32.081268 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9\": container with ID starting with 6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9 not found: ID does not exist" containerID="6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.081300 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9"} err="failed to get container status \"6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9\": rpc error: code = NotFound desc = could not find container \"6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9\": container with ID starting with 6ee324bdebe441c9924aa1d436041e2ca00c4151cd9ab9394e8d32abfa94f4d9 not found: ID does not exist" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.081318 4575 scope.go:117] "RemoveContainer" containerID="c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626" Oct 04 04:49:32 crc kubenswrapper[4575]: E1004 04:49:32.081703 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626\": container with ID starting with c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626 not found: ID does not exist" containerID="c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626" Oct 04 04:49:32 crc kubenswrapper[4575]: I1004 04:49:32.081733 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626"} err="failed to get container status \"c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626\": rpc error: code = NotFound desc = could not find container \"c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626\": container with ID starting with c2e17af1c8252e8ca83cc00f1d8cd632ad0886cfae29e38a1d5f8f94d2467626 not found: ID does not exist" Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.130793 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rxztc"] Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.131457 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rxztc" podUID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerName="registry-server" containerID="cri-o://507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177" gracePeriod=2 Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.318945 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a880646-7084-4eb9-9341-c50e28a34ec6" path="/var/lib/kubelet/pods/8a880646-7084-4eb9-9341-c50e28a34ec6/volumes" Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.608399 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.657688 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-catalog-content\") pod \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.657765 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgfg6\" (UniqueName: \"kubernetes.io/projected/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-kube-api-access-fgfg6\") pod \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.657844 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-utilities\") pod \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\" (UID: \"27c6d7db-7681-4d05-adde-48e6fb2d4eb9\") " Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.665264 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-kube-api-access-fgfg6" (OuterVolumeSpecName: "kube-api-access-fgfg6") pod "27c6d7db-7681-4d05-adde-48e6fb2d4eb9" (UID: "27c6d7db-7681-4d05-adde-48e6fb2d4eb9"). InnerVolumeSpecName "kube-api-access-fgfg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.665687 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-utilities" (OuterVolumeSpecName: "utilities") pod "27c6d7db-7681-4d05-adde-48e6fb2d4eb9" (UID: "27c6d7db-7681-4d05-adde-48e6fb2d4eb9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.708763 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27c6d7db-7681-4d05-adde-48e6fb2d4eb9" (UID: "27c6d7db-7681-4d05-adde-48e6fb2d4eb9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.759795 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.759853 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgfg6\" (UniqueName: \"kubernetes.io/projected/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-kube-api-access-fgfg6\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:33 crc kubenswrapper[4575]: I1004 04:49:33.759866 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c6d7db-7681-4d05-adde-48e6fb2d4eb9-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.026548 4575 generic.go:334] "Generic (PLEG): container finished" podID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerID="507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177" exitCode=0 Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.026641 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxztc" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.026656 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxztc" event={"ID":"27c6d7db-7681-4d05-adde-48e6fb2d4eb9","Type":"ContainerDied","Data":"507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177"} Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.027379 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxztc" event={"ID":"27c6d7db-7681-4d05-adde-48e6fb2d4eb9","Type":"ContainerDied","Data":"2b48d93a11069007e24b62d3584d969335ee979b31d647cc1b5aac0f5b13f169"} Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.027401 4575 scope.go:117] "RemoveContainer" containerID="507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.054405 4575 scope.go:117] "RemoveContainer" containerID="6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.072750 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rxztc"] Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.077842 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rxztc"] Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.082977 4575 scope.go:117] "RemoveContainer" containerID="fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.096501 4575 scope.go:117] "RemoveContainer" containerID="507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177" Oct 04 04:49:34 crc kubenswrapper[4575]: E1004 04:49:34.099041 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177\": container with ID starting with 507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177 not found: ID does not exist" containerID="507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.099084 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177"} err="failed to get container status \"507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177\": rpc error: code = NotFound desc = could not find container \"507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177\": container with ID starting with 507e7ecd5a9d75daaacc6cba46b5e3f7850a258cad4a01781d82dfc030266177 not found: ID does not exist" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.099111 4575 scope.go:117] "RemoveContainer" containerID="6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58" Oct 04 04:49:34 crc kubenswrapper[4575]: E1004 04:49:34.100175 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58\": container with ID starting with 6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58 not found: ID does not exist" containerID="6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.100219 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58"} err="failed to get container status \"6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58\": rpc error: code = NotFound desc = could not find container \"6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58\": container with ID starting with 6ea1f9f8710b8036a2b27664c2a03eaa9ed2d672ea3a0f4b37da7bc2f40a3a58 not found: ID does not exist" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.100249 4575 scope.go:117] "RemoveContainer" containerID="fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7" Oct 04 04:49:34 crc kubenswrapper[4575]: E1004 04:49:34.100528 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7\": container with ID starting with fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7 not found: ID does not exist" containerID="fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7" Oct 04 04:49:34 crc kubenswrapper[4575]: I1004 04:49:34.100559 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7"} err="failed to get container status \"fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7\": rpc error: code = NotFound desc = could not find container \"fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7\": container with ID starting with fc62343c763785a51f0a7f13c21b98b88c23e22e42d13189f70faed5f58a77f7 not found: ID does not exist" Oct 04 04:49:35 crc kubenswrapper[4575]: I1004 04:49:35.317595 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" path="/var/lib/kubelet/pods/27c6d7db-7681-4d05-adde-48e6fb2d4eb9/volumes" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.021158 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-vgclk" podUID="8bf9078b-7040-4b98-86d5-1199c124ccb1" containerName="console" containerID="cri-o://bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d" gracePeriod=15 Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.182650 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll"] Oct 04 04:49:38 crc kubenswrapper[4575]: E1004 04:49:38.183177 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerName="extract-utilities" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183189 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerName="extract-utilities" Oct 04 04:49:38 crc kubenswrapper[4575]: E1004 04:49:38.183199 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerName="extract-content" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183205 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerName="extract-content" Oct 04 04:49:38 crc kubenswrapper[4575]: E1004 04:49:38.183218 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerName="extract-content" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183224 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerName="extract-content" Oct 04 04:49:38 crc kubenswrapper[4575]: E1004 04:49:38.183232 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerName="registry-server" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183240 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerName="registry-server" Oct 04 04:49:38 crc kubenswrapper[4575]: E1004 04:49:38.183254 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerName="registry-server" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183260 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerName="registry-server" Oct 04 04:49:38 crc kubenswrapper[4575]: E1004 04:49:38.183266 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerName="extract-utilities" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183271 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerName="extract-utilities" Oct 04 04:49:38 crc kubenswrapper[4575]: E1004 04:49:38.183279 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerName="registry-server" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183284 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerName="registry-server" Oct 04 04:49:38 crc kubenswrapper[4575]: E1004 04:49:38.183293 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerName="extract-content" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183300 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerName="extract-content" Oct 04 04:49:38 crc kubenswrapper[4575]: E1004 04:49:38.183307 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerName="extract-utilities" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183313 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerName="extract-utilities" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183400 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a880646-7084-4eb9-9341-c50e28a34ec6" containerName="registry-server" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183417 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="27c6d7db-7681-4d05-adde-48e6fb2d4eb9" containerName="registry-server" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.183427 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d0921b-3c88-4f60-8aa1-051700e7605d" containerName="registry-server" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.184158 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.187532 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.201722 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll"] Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.321968 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.322050 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.322145 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m75bx\" (UniqueName: \"kubernetes.io/projected/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-kube-api-access-m75bx\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.394966 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vgclk_8bf9078b-7040-4b98-86d5-1199c124ccb1/console/0.log" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.395027 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.423235 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m75bx\" (UniqueName: \"kubernetes.io/projected/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-kube-api-access-m75bx\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.423333 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.423400 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.424001 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.424025 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.446529 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m75bx\" (UniqueName: \"kubernetes.io/projected/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-kube-api-access-m75bx\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.501438 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.524345 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-service-ca\") pod \"8bf9078b-7040-4b98-86d5-1199c124ccb1\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.524397 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-config\") pod \"8bf9078b-7040-4b98-86d5-1199c124ccb1\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.524444 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-serving-cert\") pod \"8bf9078b-7040-4b98-86d5-1199c124ccb1\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.524474 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx8vl\" (UniqueName: \"kubernetes.io/projected/8bf9078b-7040-4b98-86d5-1199c124ccb1-kube-api-access-xx8vl\") pod \"8bf9078b-7040-4b98-86d5-1199c124ccb1\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.524502 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-oauth-config\") pod \"8bf9078b-7040-4b98-86d5-1199c124ccb1\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.524541 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-oauth-serving-cert\") pod \"8bf9078b-7040-4b98-86d5-1199c124ccb1\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.524561 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-trusted-ca-bundle\") pod \"8bf9078b-7040-4b98-86d5-1199c124ccb1\" (UID: \"8bf9078b-7040-4b98-86d5-1199c124ccb1\") " Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.525250 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-config" (OuterVolumeSpecName: "console-config") pod "8bf9078b-7040-4b98-86d5-1199c124ccb1" (UID: "8bf9078b-7040-4b98-86d5-1199c124ccb1"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.525730 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "8bf9078b-7040-4b98-86d5-1199c124ccb1" (UID: "8bf9078b-7040-4b98-86d5-1199c124ccb1"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.525835 4575 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.525860 4575 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.526325 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "8bf9078b-7040-4b98-86d5-1199c124ccb1" (UID: "8bf9078b-7040-4b98-86d5-1199c124ccb1"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.527298 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-service-ca" (OuterVolumeSpecName: "service-ca") pod "8bf9078b-7040-4b98-86d5-1199c124ccb1" (UID: "8bf9078b-7040-4b98-86d5-1199c124ccb1"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.528030 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "8bf9078b-7040-4b98-86d5-1199c124ccb1" (UID: "8bf9078b-7040-4b98-86d5-1199c124ccb1"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.528453 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "8bf9078b-7040-4b98-86d5-1199c124ccb1" (UID: "8bf9078b-7040-4b98-86d5-1199c124ccb1"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.530611 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf9078b-7040-4b98-86d5-1199c124ccb1-kube-api-access-xx8vl" (OuterVolumeSpecName: "kube-api-access-xx8vl") pod "8bf9078b-7040-4b98-86d5-1199c124ccb1" (UID: "8bf9078b-7040-4b98-86d5-1199c124ccb1"). InnerVolumeSpecName "kube-api-access-xx8vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.627464 4575 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.627487 4575 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.627503 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx8vl\" (UniqueName: \"kubernetes.io/projected/8bf9078b-7040-4b98-86d5-1199c124ccb1-kube-api-access-xx8vl\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.627512 4575 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8bf9078b-7040-4b98-86d5-1199c124ccb1-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.627521 4575 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bf9078b-7040-4b98-86d5-1199c124ccb1-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:38 crc kubenswrapper[4575]: I1004 04:49:38.696216 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll"] Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.056692 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vgclk_8bf9078b-7040-4b98-86d5-1199c124ccb1/console/0.log" Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.057049 4575 generic.go:334] "Generic (PLEG): container finished" podID="8bf9078b-7040-4b98-86d5-1199c124ccb1" containerID="bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d" exitCode=2 Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.057137 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vgclk" Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.057150 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vgclk" event={"ID":"8bf9078b-7040-4b98-86d5-1199c124ccb1","Type":"ContainerDied","Data":"bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d"} Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.057260 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vgclk" event={"ID":"8bf9078b-7040-4b98-86d5-1199c124ccb1","Type":"ContainerDied","Data":"8d5595b39a358c15867d8e09b1b2d9008189812e3767a6a8e9882b0e399d82bd"} Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.057281 4575 scope.go:117] "RemoveContainer" containerID="bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d" Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.060625 4575 generic.go:334] "Generic (PLEG): container finished" podID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerID="ba5fc9349d385eb88de31d3c5c55bc357c35f3e7b1e576913863c66e2314574b" exitCode=0 Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.060672 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" event={"ID":"41fb4729-7b5d-4594-a5a6-890cf32a6ec8","Type":"ContainerDied","Data":"ba5fc9349d385eb88de31d3c5c55bc357c35f3e7b1e576913863c66e2314574b"} Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.060701 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" event={"ID":"41fb4729-7b5d-4594-a5a6-890cf32a6ec8","Type":"ContainerStarted","Data":"fefc9d42a7b8be59d6a2903e3f1263331d4531fbbb13b953a901a845c0864dbd"} Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.078135 4575 scope.go:117] "RemoveContainer" containerID="bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d" Oct 04 04:49:39 crc kubenswrapper[4575]: E1004 04:49:39.080211 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d\": container with ID starting with bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d not found: ID does not exist" containerID="bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d" Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.080239 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d"} err="failed to get container status \"bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d\": rpc error: code = NotFound desc = could not find container \"bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d\": container with ID starting with bf933e2546471c5ede323d8b2d31057e428180c5ae599cda3b32d679d0cda90d not found: ID does not exist" Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.093426 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vgclk"] Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.096781 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-vgclk"] Oct 04 04:49:39 crc kubenswrapper[4575]: I1004 04:49:39.316325 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf9078b-7040-4b98-86d5-1199c124ccb1" path="/var/lib/kubelet/pods/8bf9078b-7040-4b98-86d5-1199c124ccb1/volumes" Oct 04 04:49:41 crc kubenswrapper[4575]: I1004 04:49:41.073375 4575 generic.go:334] "Generic (PLEG): container finished" podID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerID="922a4ed83c743a75d58d99c9027910c0101c5687c039284c1fdffbea67f08513" exitCode=0 Oct 04 04:49:41 crc kubenswrapper[4575]: I1004 04:49:41.073516 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" event={"ID":"41fb4729-7b5d-4594-a5a6-890cf32a6ec8","Type":"ContainerDied","Data":"922a4ed83c743a75d58d99c9027910c0101c5687c039284c1fdffbea67f08513"} Oct 04 04:49:42 crc kubenswrapper[4575]: I1004 04:49:42.081878 4575 generic.go:334] "Generic (PLEG): container finished" podID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerID="ab826e32ba8e7fffc654bf2ce804a41d95e263a9f3a9368018701b1b47201db9" exitCode=0 Oct 04 04:49:42 crc kubenswrapper[4575]: I1004 04:49:42.082225 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" event={"ID":"41fb4729-7b5d-4594-a5a6-890cf32a6ec8","Type":"ContainerDied","Data":"ab826e32ba8e7fffc654bf2ce804a41d95e263a9f3a9368018701b1b47201db9"} Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.297822 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.396275 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-util\") pod \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.396366 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-bundle\") pod \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.396450 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m75bx\" (UniqueName: \"kubernetes.io/projected/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-kube-api-access-m75bx\") pod \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\" (UID: \"41fb4729-7b5d-4594-a5a6-890cf32a6ec8\") " Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.397434 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-bundle" (OuterVolumeSpecName: "bundle") pod "41fb4729-7b5d-4594-a5a6-890cf32a6ec8" (UID: "41fb4729-7b5d-4594-a5a6-890cf32a6ec8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.405176 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-kube-api-access-m75bx" (OuterVolumeSpecName: "kube-api-access-m75bx") pod "41fb4729-7b5d-4594-a5a6-890cf32a6ec8" (UID: "41fb4729-7b5d-4594-a5a6-890cf32a6ec8"). InnerVolumeSpecName "kube-api-access-m75bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.411455 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-util" (OuterVolumeSpecName: "util") pod "41fb4729-7b5d-4594-a5a6-890cf32a6ec8" (UID: "41fb4729-7b5d-4594-a5a6-890cf32a6ec8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.498203 4575 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-util\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.498252 4575 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:43 crc kubenswrapper[4575]: I1004 04:49:43.498266 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m75bx\" (UniqueName: \"kubernetes.io/projected/41fb4729-7b5d-4594-a5a6-890cf32a6ec8-kube-api-access-m75bx\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:44 crc kubenswrapper[4575]: I1004 04:49:44.094336 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" event={"ID":"41fb4729-7b5d-4594-a5a6-890cf32a6ec8","Type":"ContainerDied","Data":"fefc9d42a7b8be59d6a2903e3f1263331d4531fbbb13b953a901a845c0864dbd"} Oct 04 04:49:44 crc kubenswrapper[4575]: I1004 04:49:44.094374 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fefc9d42a7b8be59d6a2903e3f1263331d4531fbbb13b953a901a845c0864dbd" Oct 04 04:49:44 crc kubenswrapper[4575]: I1004 04:49:44.094430 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.208403 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5"] Oct 04 04:49:55 crc kubenswrapper[4575]: E1004 04:49:55.209963 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerName="util" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.210028 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerName="util" Oct 04 04:49:55 crc kubenswrapper[4575]: E1004 04:49:55.210084 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerName="extract" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.210145 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerName="extract" Oct 04 04:49:55 crc kubenswrapper[4575]: E1004 04:49:55.210197 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf9078b-7040-4b98-86d5-1199c124ccb1" containerName="console" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.210254 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf9078b-7040-4b98-86d5-1199c124ccb1" containerName="console" Oct 04 04:49:55 crc kubenswrapper[4575]: E1004 04:49:55.210306 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerName="pull" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.210356 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerName="pull" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.210518 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="41fb4729-7b5d-4594-a5a6-890cf32a6ec8" containerName="extract" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.210605 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf9078b-7040-4b98-86d5-1199c124ccb1" containerName="console" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.211209 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.216392 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-hwj7t" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.216417 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.216519 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.216761 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.225948 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.236933 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5"] Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.352158 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4310b719-4247-4689-991a-dffacf39c9a0-webhook-cert\") pod \"metallb-operator-controller-manager-657bc4848d-dzvq5\" (UID: \"4310b719-4247-4689-991a-dffacf39c9a0\") " pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.352202 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4310b719-4247-4689-991a-dffacf39c9a0-apiservice-cert\") pod \"metallb-operator-controller-manager-657bc4848d-dzvq5\" (UID: \"4310b719-4247-4689-991a-dffacf39c9a0\") " pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.352329 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9gcv\" (UniqueName: \"kubernetes.io/projected/4310b719-4247-4689-991a-dffacf39c9a0-kube-api-access-t9gcv\") pod \"metallb-operator-controller-manager-657bc4848d-dzvq5\" (UID: \"4310b719-4247-4689-991a-dffacf39c9a0\") " pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.453773 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9gcv\" (UniqueName: \"kubernetes.io/projected/4310b719-4247-4689-991a-dffacf39c9a0-kube-api-access-t9gcv\") pod \"metallb-operator-controller-manager-657bc4848d-dzvq5\" (UID: \"4310b719-4247-4689-991a-dffacf39c9a0\") " pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.453862 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4310b719-4247-4689-991a-dffacf39c9a0-webhook-cert\") pod \"metallb-operator-controller-manager-657bc4848d-dzvq5\" (UID: \"4310b719-4247-4689-991a-dffacf39c9a0\") " pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.453886 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4310b719-4247-4689-991a-dffacf39c9a0-apiservice-cert\") pod \"metallb-operator-controller-manager-657bc4848d-dzvq5\" (UID: \"4310b719-4247-4689-991a-dffacf39c9a0\") " pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.461507 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4310b719-4247-4689-991a-dffacf39c9a0-webhook-cert\") pod \"metallb-operator-controller-manager-657bc4848d-dzvq5\" (UID: \"4310b719-4247-4689-991a-dffacf39c9a0\") " pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.463198 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4310b719-4247-4689-991a-dffacf39c9a0-apiservice-cert\") pod \"metallb-operator-controller-manager-657bc4848d-dzvq5\" (UID: \"4310b719-4247-4689-991a-dffacf39c9a0\") " pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.472938 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9gcv\" (UniqueName: \"kubernetes.io/projected/4310b719-4247-4689-991a-dffacf39c9a0-kube-api-access-t9gcv\") pod \"metallb-operator-controller-manager-657bc4848d-dzvq5\" (UID: \"4310b719-4247-4689-991a-dffacf39c9a0\") " pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.525939 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.757744 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v"] Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.770868 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.773853 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.781687 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-jl5jp" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.781909 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.813963 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v"] Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.857399 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92hwb\" (UniqueName: \"kubernetes.io/projected/c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a-kube-api-access-92hwb\") pod \"metallb-operator-webhook-server-6d5f8fc9b4-5t69v\" (UID: \"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a\") " pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.857470 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a-webhook-cert\") pod \"metallb-operator-webhook-server-6d5f8fc9b4-5t69v\" (UID: \"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a\") " pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.857515 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a-apiservice-cert\") pod \"metallb-operator-webhook-server-6d5f8fc9b4-5t69v\" (UID: \"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a\") " pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.958730 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92hwb\" (UniqueName: \"kubernetes.io/projected/c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a-kube-api-access-92hwb\") pod \"metallb-operator-webhook-server-6d5f8fc9b4-5t69v\" (UID: \"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a\") " pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.958807 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a-webhook-cert\") pod \"metallb-operator-webhook-server-6d5f8fc9b4-5t69v\" (UID: \"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a\") " pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.958847 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a-apiservice-cert\") pod \"metallb-operator-webhook-server-6d5f8fc9b4-5t69v\" (UID: \"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a\") " pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.967996 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a-webhook-cert\") pod \"metallb-operator-webhook-server-6d5f8fc9b4-5t69v\" (UID: \"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a\") " pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.977116 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a-apiservice-cert\") pod \"metallb-operator-webhook-server-6d5f8fc9b4-5t69v\" (UID: \"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a\") " pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:55 crc kubenswrapper[4575]: I1004 04:49:55.996782 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5"] Oct 04 04:49:56 crc kubenswrapper[4575]: I1004 04:49:56.030451 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92hwb\" (UniqueName: \"kubernetes.io/projected/c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a-kube-api-access-92hwb\") pod \"metallb-operator-webhook-server-6d5f8fc9b4-5t69v\" (UID: \"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a\") " pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:56 crc kubenswrapper[4575]: I1004 04:49:56.098232 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:49:56 crc kubenswrapper[4575]: I1004 04:49:56.167114 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" event={"ID":"4310b719-4247-4689-991a-dffacf39c9a0","Type":"ContainerStarted","Data":"34921a548aeb828cf278d4b6de3cb32ce67f542f641f58ba79399124afd123c7"} Oct 04 04:49:56 crc kubenswrapper[4575]: I1004 04:49:56.410559 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v"] Oct 04 04:49:56 crc kubenswrapper[4575]: W1004 04:49:56.419074 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5cbf36b_ad72_4ae9_adeb_3d1a206a4d8a.slice/crio-c22ea88c4fdae7a67c18449b59704bb5ed646920108b5eae1575d55f738b84a2 WatchSource:0}: Error finding container c22ea88c4fdae7a67c18449b59704bb5ed646920108b5eae1575d55f738b84a2: Status 404 returned error can't find the container with id c22ea88c4fdae7a67c18449b59704bb5ed646920108b5eae1575d55f738b84a2 Oct 04 04:49:57 crc kubenswrapper[4575]: I1004 04:49:57.175928 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" event={"ID":"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a","Type":"ContainerStarted","Data":"c22ea88c4fdae7a67c18449b59704bb5ed646920108b5eae1575d55f738b84a2"} Oct 04 04:50:06 crc kubenswrapper[4575]: I1004 04:50:06.237777 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" event={"ID":"4310b719-4247-4689-991a-dffacf39c9a0","Type":"ContainerStarted","Data":"300336a486e3b908bd40c9eaf7e77f7954b4b507e5213e174669b6f95682693d"} Oct 04 04:50:06 crc kubenswrapper[4575]: I1004 04:50:06.238400 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:50:06 crc kubenswrapper[4575]: I1004 04:50:06.239472 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" event={"ID":"c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a","Type":"ContainerStarted","Data":"c579b3c34f9c61a75ab9c25512e22a6be10d850e40b5d727134f7ab2dc467278"} Oct 04 04:50:06 crc kubenswrapper[4575]: I1004 04:50:06.239563 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:50:06 crc kubenswrapper[4575]: I1004 04:50:06.259347 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" podStartSLOduration=2.115939499 podStartE2EDuration="11.259328612s" podCreationTimestamp="2025-10-04 04:49:55 +0000 UTC" firstStartedPulling="2025-10-04 04:49:55.981872902 +0000 UTC m=+947.310431716" lastFinishedPulling="2025-10-04 04:50:05.125262015 +0000 UTC m=+956.453820829" observedRunningTime="2025-10-04 04:50:06.25785956 +0000 UTC m=+957.586418384" watchObservedRunningTime="2025-10-04 04:50:06.259328612 +0000 UTC m=+957.587887426" Oct 04 04:50:06 crc kubenswrapper[4575]: I1004 04:50:06.279062 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" podStartSLOduration=2.556936054 podStartE2EDuration="11.279043278s" podCreationTimestamp="2025-10-04 04:49:55 +0000 UTC" firstStartedPulling="2025-10-04 04:49:56.422648721 +0000 UTC m=+947.751207535" lastFinishedPulling="2025-10-04 04:50:05.144755945 +0000 UTC m=+956.473314759" observedRunningTime="2025-10-04 04:50:06.278105871 +0000 UTC m=+957.606664705" watchObservedRunningTime="2025-10-04 04:50:06.279043278 +0000 UTC m=+957.607602102" Oct 04 04:50:08 crc kubenswrapper[4575]: I1004 04:50:08.446769 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:50:08 crc kubenswrapper[4575]: I1004 04:50:08.446842 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:50:16 crc kubenswrapper[4575]: I1004 04:50:16.102723 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6d5f8fc9b4-5t69v" Oct 04 04:50:35 crc kubenswrapper[4575]: I1004 04:50:35.529599 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-657bc4848d-dzvq5" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.361349 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-9w4x4"] Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.364084 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.372199 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr"] Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.373149 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.373502 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.373660 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-sfkxq" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.377358 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.377616 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.450496 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn5gq\" (UniqueName: \"kubernetes.io/projected/ea7f8ebe-6b41-4333-9213-8aa2f6180f73-kube-api-access-vn5gq\") pod \"frr-k8s-webhook-server-64bf5d555-q5zvr\" (UID: \"ea7f8ebe-6b41-4333-9213-8aa2f6180f73\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.450617 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-frr-conf\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.450664 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-metrics\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.450697 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cmbs\" (UniqueName: \"kubernetes.io/projected/27a10d08-b3eb-4d6d-8815-9db3a56313c5-kube-api-access-8cmbs\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.450730 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27a10d08-b3eb-4d6d-8815-9db3a56313c5-metrics-certs\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.450810 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/27a10d08-b3eb-4d6d-8815-9db3a56313c5-frr-startup\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.450832 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-frr-sockets\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.450898 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea7f8ebe-6b41-4333-9213-8aa2f6180f73-cert\") pod \"frr-k8s-webhook-server-64bf5d555-q5zvr\" (UID: \"ea7f8ebe-6b41-4333-9213-8aa2f6180f73\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.450954 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-reloader\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.454521 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr"] Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.552754 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-frr-conf\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.552815 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-metrics\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.552848 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cmbs\" (UniqueName: \"kubernetes.io/projected/27a10d08-b3eb-4d6d-8815-9db3a56313c5-kube-api-access-8cmbs\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.552875 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27a10d08-b3eb-4d6d-8815-9db3a56313c5-metrics-certs\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.552932 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/27a10d08-b3eb-4d6d-8815-9db3a56313c5-frr-startup\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.552954 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-frr-sockets\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.552992 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea7f8ebe-6b41-4333-9213-8aa2f6180f73-cert\") pod \"frr-k8s-webhook-server-64bf5d555-q5zvr\" (UID: \"ea7f8ebe-6b41-4333-9213-8aa2f6180f73\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.553015 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-reloader\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.553039 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn5gq\" (UniqueName: \"kubernetes.io/projected/ea7f8ebe-6b41-4333-9213-8aa2f6180f73-kube-api-access-vn5gq\") pod \"frr-k8s-webhook-server-64bf5d555-q5zvr\" (UID: \"ea7f8ebe-6b41-4333-9213-8aa2f6180f73\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.553527 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-frr-conf\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.553786 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-metrics\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.554040 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-frr-sockets\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: E1004 04:50:36.554131 4575 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 04 04:50:36 crc kubenswrapper[4575]: E1004 04:50:36.554192 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27a10d08-b3eb-4d6d-8815-9db3a56313c5-metrics-certs podName:27a10d08-b3eb-4d6d-8815-9db3a56313c5 nodeName:}" failed. No retries permitted until 2025-10-04 04:50:37.054174886 +0000 UTC m=+988.382733710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27a10d08-b3eb-4d6d-8815-9db3a56313c5-metrics-certs") pod "frr-k8s-9w4x4" (UID: "27a10d08-b3eb-4d6d-8815-9db3a56313c5") : secret "frr-k8s-certs-secret" not found Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.554307 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/27a10d08-b3eb-4d6d-8815-9db3a56313c5-reloader\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: E1004 04:50:36.554498 4575 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 04 04:50:36 crc kubenswrapper[4575]: E1004 04:50:36.554676 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ea7f8ebe-6b41-4333-9213-8aa2f6180f73-cert podName:ea7f8ebe-6b41-4333-9213-8aa2f6180f73 nodeName:}" failed. No retries permitted until 2025-10-04 04:50:37.054652799 +0000 UTC m=+988.383211613 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ea7f8ebe-6b41-4333-9213-8aa2f6180f73-cert") pod "frr-k8s-webhook-server-64bf5d555-q5zvr" (UID: "ea7f8ebe-6b41-4333-9213-8aa2f6180f73") : secret "frr-k8s-webhook-server-cert" not found Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.554711 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/27a10d08-b3eb-4d6d-8815-9db3a56313c5-frr-startup\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.573643 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn5gq\" (UniqueName: \"kubernetes.io/projected/ea7f8ebe-6b41-4333-9213-8aa2f6180f73-kube-api-access-vn5gq\") pod \"frr-k8s-webhook-server-64bf5d555-q5zvr\" (UID: \"ea7f8ebe-6b41-4333-9213-8aa2f6180f73\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.578155 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cmbs\" (UniqueName: \"kubernetes.io/projected/27a10d08-b3eb-4d6d-8815-9db3a56313c5-kube-api-access-8cmbs\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.594770 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-mm46z"] Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.595927 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: W1004 04:50:36.603325 4575 reflector.go:561] object-"metallb-system"/"speaker-dockercfg-t9pmw": failed to list *v1.Secret: secrets "speaker-dockercfg-t9pmw" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 04 04:50:36 crc kubenswrapper[4575]: E1004 04:50:36.603596 4575 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"speaker-dockercfg-t9pmw\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"speaker-dockercfg-t9pmw\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:50:36 crc kubenswrapper[4575]: W1004 04:50:36.603521 4575 reflector.go:561] object-"metallb-system"/"metallb-memberlist": failed to list *v1.Secret: secrets "metallb-memberlist" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 04 04:50:36 crc kubenswrapper[4575]: E1004 04:50:36.603863 4575 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-memberlist\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-memberlist\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:50:36 crc kubenswrapper[4575]: W1004 04:50:36.604868 4575 reflector.go:561] object-"metallb-system"/"metallb-excludel2": failed to list *v1.ConfigMap: configmaps "metallb-excludel2" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 04 04:50:36 crc kubenswrapper[4575]: E1004 04:50:36.604902 4575 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-excludel2\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"metallb-excludel2\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.645487 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.654426 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-7ztmk"] Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.655730 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.662453 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.679877 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-7ztmk"] Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.755772 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ba9c8a20-378b-44b0-8db7-78d5ced7671e-metrics-certs\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.756082 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl6vl\" (UniqueName: \"kubernetes.io/projected/ba9c8a20-378b-44b0-8db7-78d5ced7671e-kube-api-access-vl6vl\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.756205 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/eb053b2d-c0d8-456f-9437-ab93727a1900-memberlist\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.756306 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hrrs\" (UniqueName: \"kubernetes.io/projected/eb053b2d-c0d8-456f-9437-ab93727a1900-kube-api-access-6hrrs\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.756415 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb053b2d-c0d8-456f-9437-ab93727a1900-metrics-certs\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.756527 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ba9c8a20-378b-44b0-8db7-78d5ced7671e-cert\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.756720 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/eb053b2d-c0d8-456f-9437-ab93727a1900-metallb-excludel2\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.857899 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/eb053b2d-c0d8-456f-9437-ab93727a1900-metallb-excludel2\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.858213 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ba9c8a20-378b-44b0-8db7-78d5ced7671e-metrics-certs\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.858333 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl6vl\" (UniqueName: \"kubernetes.io/projected/ba9c8a20-378b-44b0-8db7-78d5ced7671e-kube-api-access-vl6vl\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.858450 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/eb053b2d-c0d8-456f-9437-ab93727a1900-memberlist\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.858555 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hrrs\" (UniqueName: \"kubernetes.io/projected/eb053b2d-c0d8-456f-9437-ab93727a1900-kube-api-access-6hrrs\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: E1004 04:50:36.858380 4575 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 04 04:50:36 crc kubenswrapper[4575]: E1004 04:50:36.858761 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba9c8a20-378b-44b0-8db7-78d5ced7671e-metrics-certs podName:ba9c8a20-378b-44b0-8db7-78d5ced7671e nodeName:}" failed. No retries permitted until 2025-10-04 04:50:37.358739825 +0000 UTC m=+988.687298639 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ba9c8a20-378b-44b0-8db7-78d5ced7671e-metrics-certs") pod "controller-68d546b9d8-7ztmk" (UID: "ba9c8a20-378b-44b0-8db7-78d5ced7671e") : secret "controller-certs-secret" not found Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.858695 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb053b2d-c0d8-456f-9437-ab93727a1900-metrics-certs\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.858875 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ba9c8a20-378b-44b0-8db7-78d5ced7671e-cert\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.861486 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.864849 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb053b2d-c0d8-456f-9437-ab93727a1900-metrics-certs\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.873987 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ba9c8a20-378b-44b0-8db7-78d5ced7671e-cert\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.878652 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hrrs\" (UniqueName: \"kubernetes.io/projected/eb053b2d-c0d8-456f-9437-ab93727a1900-kube-api-access-6hrrs\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:36 crc kubenswrapper[4575]: I1004 04:50:36.882435 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl6vl\" (UniqueName: \"kubernetes.io/projected/ba9c8a20-378b-44b0-8db7-78d5ced7671e-kube-api-access-vl6vl\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.061646 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea7f8ebe-6b41-4333-9213-8aa2f6180f73-cert\") pod \"frr-k8s-webhook-server-64bf5d555-q5zvr\" (UID: \"ea7f8ebe-6b41-4333-9213-8aa2f6180f73\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.061771 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27a10d08-b3eb-4d6d-8815-9db3a56313c5-metrics-certs\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.065553 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27a10d08-b3eb-4d6d-8815-9db3a56313c5-metrics-certs\") pod \"frr-k8s-9w4x4\" (UID: \"27a10d08-b3eb-4d6d-8815-9db3a56313c5\") " pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.067124 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea7f8ebe-6b41-4333-9213-8aa2f6180f73-cert\") pod \"frr-k8s-webhook-server-64bf5d555-q5zvr\" (UID: \"ea7f8ebe-6b41-4333-9213-8aa2f6180f73\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.289906 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.301782 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.366034 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ba9c8a20-378b-44b0-8db7-78d5ced7671e-metrics-certs\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.372644 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ba9c8a20-378b-44b0-8db7-78d5ced7671e-metrics-certs\") pod \"controller-68d546b9d8-7ztmk\" (UID: \"ba9c8a20-378b-44b0-8db7-78d5ced7671e\") " pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.484907 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-t9pmw" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.570665 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.804163 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr"] Oct 04 04:50:37 crc kubenswrapper[4575]: W1004 04:50:37.832835 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea7f8ebe_6b41_4333_9213_8aa2f6180f73.slice/crio-a909d570555a0aaaf0fe2b6c7886c1a7f798ba22e5ce30613d9c05b7bfe70ab2 WatchSource:0}: Error finding container a909d570555a0aaaf0fe2b6c7886c1a7f798ba22e5ce30613d9c05b7bfe70ab2: Status 404 returned error can't find the container with id a909d570555a0aaaf0fe2b6c7886c1a7f798ba22e5ce30613d9c05b7bfe70ab2 Oct 04 04:50:37 crc kubenswrapper[4575]: E1004 04:50:37.858465 4575 configmap.go:193] Couldn't get configMap metallb-system/metallb-excludel2: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:50:37 crc kubenswrapper[4575]: E1004 04:50:37.858577 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/eb053b2d-c0d8-456f-9437-ab93727a1900-metallb-excludel2 podName:eb053b2d-c0d8-456f-9437-ab93727a1900 nodeName:}" failed. No retries permitted until 2025-10-04 04:50:38.358553638 +0000 UTC m=+989.687112452 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metallb-excludel2" (UniqueName: "kubernetes.io/configmap/eb053b2d-c0d8-456f-9437-ab93727a1900-metallb-excludel2") pod "speaker-mm46z" (UID: "eb053b2d-c0d8-456f-9437-ab93727a1900") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:50:37 crc kubenswrapper[4575]: E1004 04:50:37.859120 4575 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: failed to sync secret cache: timed out waiting for the condition Oct 04 04:50:37 crc kubenswrapper[4575]: E1004 04:50:37.859332 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb053b2d-c0d8-456f-9437-ab93727a1900-memberlist podName:eb053b2d-c0d8-456f-9437-ab93727a1900 nodeName:}" failed. No retries permitted until 2025-10-04 04:50:38.359319751 +0000 UTC m=+989.687878565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/eb053b2d-c0d8-456f-9437-ab93727a1900-memberlist") pod "speaker-mm46z" (UID: "eb053b2d-c0d8-456f-9437-ab93727a1900") : failed to sync secret cache: timed out waiting for the condition Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.893862 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-7ztmk"] Oct 04 04:50:37 crc kubenswrapper[4575]: W1004 04:50:37.902651 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba9c8a20_378b_44b0_8db7_78d5ced7671e.slice/crio-db4dab7c48c0d409517e504cadfddc0749de210433e6a845fa0d38d3d25a0011 WatchSource:0}: Error finding container db4dab7c48c0d409517e504cadfddc0749de210433e6a845fa0d38d3d25a0011: Status 404 returned error can't find the container with id db4dab7c48c0d409517e504cadfddc0749de210433e6a845fa0d38d3d25a0011 Oct 04 04:50:37 crc kubenswrapper[4575]: I1004 04:50:37.959446 4575 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.191538 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.380041 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/eb053b2d-c0d8-456f-9437-ab93727a1900-metallb-excludel2\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.380131 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/eb053b2d-c0d8-456f-9437-ab93727a1900-memberlist\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.380977 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/eb053b2d-c0d8-456f-9437-ab93727a1900-metallb-excludel2\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.390239 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/eb053b2d-c0d8-456f-9437-ab93727a1900-memberlist\") pod \"speaker-mm46z\" (UID: \"eb053b2d-c0d8-456f-9437-ab93727a1900\") " pod="metallb-system/speaker-mm46z" Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.422348 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" event={"ID":"ea7f8ebe-6b41-4333-9213-8aa2f6180f73","Type":"ContainerStarted","Data":"a909d570555a0aaaf0fe2b6c7886c1a7f798ba22e5ce30613d9c05b7bfe70ab2"} Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.423697 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerStarted","Data":"b202911034adfd013a0d3624ba2176ef9df594f506d6050170296f0ce114691d"} Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.426433 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-7ztmk" event={"ID":"ba9c8a20-378b-44b0-8db7-78d5ced7671e","Type":"ContainerStarted","Data":"f791d1a031cef22107f7aa75145ac0767bdd8754af7d0a348d1444b1e4b225b1"} Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.426487 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-7ztmk" event={"ID":"ba9c8a20-378b-44b0-8db7-78d5ced7671e","Type":"ContainerStarted","Data":"5da545ce8e62d991b8c9d60aacc5da66185c769385d88c1f90ebb8d4d929a215"} Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.426520 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-7ztmk" event={"ID":"ba9c8a20-378b-44b0-8db7-78d5ced7671e","Type":"ContainerStarted","Data":"db4dab7c48c0d409517e504cadfddc0749de210433e6a845fa0d38d3d25a0011"} Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.426717 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.432453 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mm46z" Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.450127 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.450186 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:50:38 crc kubenswrapper[4575]: I1004 04:50:38.458430 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-7ztmk" podStartSLOduration=2.458401903 podStartE2EDuration="2.458401903s" podCreationTimestamp="2025-10-04 04:50:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:50:38.450032593 +0000 UTC m=+989.778591417" watchObservedRunningTime="2025-10-04 04:50:38.458401903 +0000 UTC m=+989.786960717" Oct 04 04:50:38 crc kubenswrapper[4575]: W1004 04:50:38.469315 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb053b2d_c0d8_456f_9437_ab93727a1900.slice/crio-f55139a01b2c352defab02ad43b05494006c9fd7861aa710f66cd2fc13ff07bd WatchSource:0}: Error finding container f55139a01b2c352defab02ad43b05494006c9fd7861aa710f66cd2fc13ff07bd: Status 404 returned error can't find the container with id f55139a01b2c352defab02ad43b05494006c9fd7861aa710f66cd2fc13ff07bd Oct 04 04:50:39 crc kubenswrapper[4575]: I1004 04:50:39.448625 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mm46z" event={"ID":"eb053b2d-c0d8-456f-9437-ab93727a1900","Type":"ContainerStarted","Data":"b31d05f4b5c5ebb07fb3524ef3e7f1c222f2d79de6664aa7d73552d37ee24367"} Oct 04 04:50:39 crc kubenswrapper[4575]: I1004 04:50:39.448921 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mm46z" event={"ID":"eb053b2d-c0d8-456f-9437-ab93727a1900","Type":"ContainerStarted","Data":"58db20ad4123f29dfa469d538a9f914bec37b340ce9b12be3c74c905235615a3"} Oct 04 04:50:39 crc kubenswrapper[4575]: I1004 04:50:39.448937 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mm46z" event={"ID":"eb053b2d-c0d8-456f-9437-ab93727a1900","Type":"ContainerStarted","Data":"f55139a01b2c352defab02ad43b05494006c9fd7861aa710f66cd2fc13ff07bd"} Oct 04 04:50:39 crc kubenswrapper[4575]: I1004 04:50:39.449732 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-mm46z" Oct 04 04:50:39 crc kubenswrapper[4575]: I1004 04:50:39.475495 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-mm46z" podStartSLOduration=3.475476932 podStartE2EDuration="3.475476932s" podCreationTimestamp="2025-10-04 04:50:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:50:39.474936356 +0000 UTC m=+990.803495190" watchObservedRunningTime="2025-10-04 04:50:39.475476932 +0000 UTC m=+990.804035746" Oct 04 04:50:48 crc kubenswrapper[4575]: I1004 04:50:48.436994 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-mm46z" Oct 04 04:50:50 crc kubenswrapper[4575]: I1004 04:50:50.527791 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" event={"ID":"ea7f8ebe-6b41-4333-9213-8aa2f6180f73","Type":"ContainerStarted","Data":"5661aee4e0bf15a3c7b09689a91dd6adf8f98efec0c988c4934e8e24b901eeac"} Oct 04 04:50:50 crc kubenswrapper[4575]: I1004 04:50:50.528146 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:50:50 crc kubenswrapper[4575]: I1004 04:50:50.529866 4575 generic.go:334] "Generic (PLEG): container finished" podID="27a10d08-b3eb-4d6d-8815-9db3a56313c5" containerID="57cc6b114e709cf19578a3a56b3b5b0fedea6c091265d70cd362f007d8bd9a48" exitCode=0 Oct 04 04:50:50 crc kubenswrapper[4575]: I1004 04:50:50.529903 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerDied","Data":"57cc6b114e709cf19578a3a56b3b5b0fedea6c091265d70cd362f007d8bd9a48"} Oct 04 04:50:50 crc kubenswrapper[4575]: I1004 04:50:50.581803 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" podStartSLOduration=2.783010829 podStartE2EDuration="14.581781424s" podCreationTimestamp="2025-10-04 04:50:36 +0000 UTC" firstStartedPulling="2025-10-04 04:50:37.840028797 +0000 UTC m=+989.168587611" lastFinishedPulling="2025-10-04 04:50:49.638799392 +0000 UTC m=+1000.967358206" observedRunningTime="2025-10-04 04:50:50.550491696 +0000 UTC m=+1001.879050520" watchObservedRunningTime="2025-10-04 04:50:50.581781424 +0000 UTC m=+1001.910340238" Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.283530 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-ttqtk"] Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.284822 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ttqtk" Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.296170 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-fcqgj" Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.296527 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.298267 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.307685 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ttqtk"] Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.401383 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqzb4\" (UniqueName: \"kubernetes.io/projected/c3cd2298-f99d-4640-904d-d701c473347b-kube-api-access-xqzb4\") pod \"openstack-operator-index-ttqtk\" (UID: \"c3cd2298-f99d-4640-904d-d701c473347b\") " pod="openstack-operators/openstack-operator-index-ttqtk" Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.503273 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqzb4\" (UniqueName: \"kubernetes.io/projected/c3cd2298-f99d-4640-904d-d701c473347b-kube-api-access-xqzb4\") pod \"openstack-operator-index-ttqtk\" (UID: \"c3cd2298-f99d-4640-904d-d701c473347b\") " pod="openstack-operators/openstack-operator-index-ttqtk" Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.521090 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqzb4\" (UniqueName: \"kubernetes.io/projected/c3cd2298-f99d-4640-904d-d701c473347b-kube-api-access-xqzb4\") pod \"openstack-operator-index-ttqtk\" (UID: \"c3cd2298-f99d-4640-904d-d701c473347b\") " pod="openstack-operators/openstack-operator-index-ttqtk" Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.537025 4575 generic.go:334] "Generic (PLEG): container finished" podID="27a10d08-b3eb-4d6d-8815-9db3a56313c5" containerID="526c2586d9def742593c664457cb48f9b47381ae837cc31b44e089bbf8d05233" exitCode=0 Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.537128 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerDied","Data":"526c2586d9def742593c664457cb48f9b47381ae837cc31b44e089bbf8d05233"} Oct 04 04:50:51 crc kubenswrapper[4575]: I1004 04:50:51.606545 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ttqtk" Oct 04 04:50:52 crc kubenswrapper[4575]: I1004 04:50:52.097875 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ttqtk"] Oct 04 04:50:52 crc kubenswrapper[4575]: I1004 04:50:52.543095 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ttqtk" event={"ID":"c3cd2298-f99d-4640-904d-d701c473347b","Type":"ContainerStarted","Data":"ab4a4b8efb5c189e1515d49c13a71ddec34672a9b8a2f15dd42caaf4ea4c151f"} Oct 04 04:50:53 crc kubenswrapper[4575]: I1004 04:50:53.551231 4575 generic.go:334] "Generic (PLEG): container finished" podID="27a10d08-b3eb-4d6d-8815-9db3a56313c5" containerID="9101bd611b5a03327ef935283db14ca3d7ed3f321a79a6994e0a4f37a6f7af61" exitCode=0 Oct 04 04:50:53 crc kubenswrapper[4575]: I1004 04:50:53.551281 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerDied","Data":"9101bd611b5a03327ef935283db14ca3d7ed3f321a79a6994e0a4f37a6f7af61"} Oct 04 04:50:54 crc kubenswrapper[4575]: I1004 04:50:54.459538 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-ttqtk"] Oct 04 04:50:54 crc kubenswrapper[4575]: I1004 04:50:54.566092 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerStarted","Data":"35b3b43a16a42d9e481be47d3cb9b3237b154ffee7717b4ddb1b7f5621e81d3f"} Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.064739 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-d6k87"] Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.066380 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-d6k87" Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.076852 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-d6k87"] Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.158582 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvckj\" (UniqueName: \"kubernetes.io/projected/3080dcb3-ddf0-4da4-956a-e90af6fef1b5-kube-api-access-rvckj\") pod \"openstack-operator-index-d6k87\" (UID: \"3080dcb3-ddf0-4da4-956a-e90af6fef1b5\") " pod="openstack-operators/openstack-operator-index-d6k87" Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.259962 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvckj\" (UniqueName: \"kubernetes.io/projected/3080dcb3-ddf0-4da4-956a-e90af6fef1b5-kube-api-access-rvckj\") pod \"openstack-operator-index-d6k87\" (UID: \"3080dcb3-ddf0-4da4-956a-e90af6fef1b5\") " pod="openstack-operators/openstack-operator-index-d6k87" Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.348471 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvckj\" (UniqueName: \"kubernetes.io/projected/3080dcb3-ddf0-4da4-956a-e90af6fef1b5-kube-api-access-rvckj\") pod \"openstack-operator-index-d6k87\" (UID: \"3080dcb3-ddf0-4da4-956a-e90af6fef1b5\") " pod="openstack-operators/openstack-operator-index-d6k87" Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.384089 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-d6k87" Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.584988 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerStarted","Data":"836283c3512a03259fcd9145fea88efa5a7e016d770f70add134e0cf057a9267"} Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.586092 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerStarted","Data":"51eb3384cfc9be3b17e0266db0550cfe3c9581ab1d21209555d113c12d315498"} Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.586108 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerStarted","Data":"53576f93f7328ff0b7665a11ecc4c264547bbe546d5ebae8a08970ce2ae7079f"} Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.586117 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerStarted","Data":"43cefba63ee4c2444fc698a811be42ac269a9d2f6eac60ab53f5366b87491edb"} Oct 04 04:50:55 crc kubenswrapper[4575]: I1004 04:50:55.823990 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-d6k87"] Oct 04 04:50:56 crc kubenswrapper[4575]: I1004 04:50:56.631727 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9w4x4" event={"ID":"27a10d08-b3eb-4d6d-8815-9db3a56313c5","Type":"ContainerStarted","Data":"c3d80f30259c335c44510726449a3eead78ff404d7eb72de64079d3b211ed24c"} Oct 04 04:50:56 crc kubenswrapper[4575]: I1004 04:50:56.634410 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-d6k87" event={"ID":"3080dcb3-ddf0-4da4-956a-e90af6fef1b5","Type":"ContainerStarted","Data":"eac47f38671ce03fddd85c0d102626050e0bd04baab70b782f761add53cd0fe4"} Oct 04 04:50:56 crc kubenswrapper[4575]: I1004 04:50:56.634549 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:56 crc kubenswrapper[4575]: I1004 04:50:56.660459 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-9w4x4" podStartSLOduration=8.48667647 podStartE2EDuration="20.660441707s" podCreationTimestamp="2025-10-04 04:50:36 +0000 UTC" firstStartedPulling="2025-10-04 04:50:37.437114724 +0000 UTC m=+988.765673538" lastFinishedPulling="2025-10-04 04:50:49.610879951 +0000 UTC m=+1000.939438775" observedRunningTime="2025-10-04 04:50:56.65912926 +0000 UTC m=+1007.987688084" watchObservedRunningTime="2025-10-04 04:50:56.660441707 +0000 UTC m=+1007.989000521" Oct 04 04:50:57 crc kubenswrapper[4575]: I1004 04:50:57.290643 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:57 crc kubenswrapper[4575]: I1004 04:50:57.334136 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:50:57 crc kubenswrapper[4575]: I1004 04:50:57.575536 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-7ztmk" Oct 04 04:51:06 crc kubenswrapper[4575]: I1004 04:51:06.700205 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ttqtk" event={"ID":"c3cd2298-f99d-4640-904d-d701c473347b","Type":"ContainerStarted","Data":"0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c"} Oct 04 04:51:06 crc kubenswrapper[4575]: I1004 04:51:06.700318 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-ttqtk" podUID="c3cd2298-f99d-4640-904d-d701c473347b" containerName="registry-server" containerID="cri-o://0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c" gracePeriod=2 Oct 04 04:51:06 crc kubenswrapper[4575]: I1004 04:51:06.701215 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-d6k87" event={"ID":"3080dcb3-ddf0-4da4-956a-e90af6fef1b5","Type":"ContainerStarted","Data":"e0f3c74bfe3507deeeed33a2184feca4344f6e8549c9878393b87aade88b4a01"} Oct 04 04:51:06 crc kubenswrapper[4575]: I1004 04:51:06.719485 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-ttqtk" podStartSLOduration=1.713332995 podStartE2EDuration="15.719464121s" podCreationTimestamp="2025-10-04 04:50:51 +0000 UTC" firstStartedPulling="2025-10-04 04:50:52.108395444 +0000 UTC m=+1003.436954258" lastFinishedPulling="2025-10-04 04:51:06.11452657 +0000 UTC m=+1017.443085384" observedRunningTime="2025-10-04 04:51:06.714994022 +0000 UTC m=+1018.043552846" watchObservedRunningTime="2025-10-04 04:51:06.719464121 +0000 UTC m=+1018.048022955" Oct 04 04:51:06 crc kubenswrapper[4575]: I1004 04:51:06.727779 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-d6k87" podStartSLOduration=1.049565757 podStartE2EDuration="11.727762119s" podCreationTimestamp="2025-10-04 04:50:55 +0000 UTC" firstStartedPulling="2025-10-04 04:50:55.841018472 +0000 UTC m=+1007.169577286" lastFinishedPulling="2025-10-04 04:51:06.519214834 +0000 UTC m=+1017.847773648" observedRunningTime="2025-10-04 04:51:06.726981136 +0000 UTC m=+1018.055539970" watchObservedRunningTime="2025-10-04 04:51:06.727762119 +0000 UTC m=+1018.056320923" Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.084422 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ttqtk" Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.230127 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqzb4\" (UniqueName: \"kubernetes.io/projected/c3cd2298-f99d-4640-904d-d701c473347b-kube-api-access-xqzb4\") pod \"c3cd2298-f99d-4640-904d-d701c473347b\" (UID: \"c3cd2298-f99d-4640-904d-d701c473347b\") " Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.235782 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3cd2298-f99d-4640-904d-d701c473347b-kube-api-access-xqzb4" (OuterVolumeSpecName: "kube-api-access-xqzb4") pod "c3cd2298-f99d-4640-904d-d701c473347b" (UID: "c3cd2298-f99d-4640-904d-d701c473347b"). InnerVolumeSpecName "kube-api-access-xqzb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.292670 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-9w4x4" Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.307954 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-q5zvr" Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.333729 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqzb4\" (UniqueName: \"kubernetes.io/projected/c3cd2298-f99d-4640-904d-d701c473347b-kube-api-access-xqzb4\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.709387 4575 generic.go:334] "Generic (PLEG): container finished" podID="c3cd2298-f99d-4640-904d-d701c473347b" containerID="0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c" exitCode=0 Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.709473 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ttqtk" Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.709476 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ttqtk" event={"ID":"c3cd2298-f99d-4640-904d-d701c473347b","Type":"ContainerDied","Data":"0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c"} Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.709567 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ttqtk" event={"ID":"c3cd2298-f99d-4640-904d-d701c473347b","Type":"ContainerDied","Data":"ab4a4b8efb5c189e1515d49c13a71ddec34672a9b8a2f15dd42caaf4ea4c151f"} Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.709619 4575 scope.go:117] "RemoveContainer" containerID="0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c" Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.728614 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-ttqtk"] Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.732139 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-ttqtk"] Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.732836 4575 scope.go:117] "RemoveContainer" containerID="0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c" Oct 04 04:51:07 crc kubenswrapper[4575]: E1004 04:51:07.733384 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c\": container with ID starting with 0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c not found: ID does not exist" containerID="0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c" Oct 04 04:51:07 crc kubenswrapper[4575]: I1004 04:51:07.733485 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c"} err="failed to get container status \"0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c\": rpc error: code = NotFound desc = could not find container \"0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c\": container with ID starting with 0d5658f0bffb86c00aa493e8a4edcc065acaad811c54e60ac5f71301f8652a9c not found: ID does not exist" Oct 04 04:51:08 crc kubenswrapper[4575]: I1004 04:51:08.446631 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:51:08 crc kubenswrapper[4575]: I1004 04:51:08.446694 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:51:08 crc kubenswrapper[4575]: I1004 04:51:08.446739 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:51:08 crc kubenswrapper[4575]: I1004 04:51:08.447437 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6585e9613991e3a385e6fad309ba0255360c49e2f3c2664f8dc4f993639cb30"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:51:08 crc kubenswrapper[4575]: I1004 04:51:08.447501 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://e6585e9613991e3a385e6fad309ba0255360c49e2f3c2664f8dc4f993639cb30" gracePeriod=600 Oct 04 04:51:09 crc kubenswrapper[4575]: I1004 04:51:09.325938 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3cd2298-f99d-4640-904d-d701c473347b" path="/var/lib/kubelet/pods/c3cd2298-f99d-4640-904d-d701c473347b/volumes" Oct 04 04:51:09 crc kubenswrapper[4575]: I1004 04:51:09.727272 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="e6585e9613991e3a385e6fad309ba0255360c49e2f3c2664f8dc4f993639cb30" exitCode=0 Oct 04 04:51:09 crc kubenswrapper[4575]: I1004 04:51:09.727401 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"e6585e9613991e3a385e6fad309ba0255360c49e2f3c2664f8dc4f993639cb30"} Oct 04 04:51:09 crc kubenswrapper[4575]: I1004 04:51:09.727777 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"3ed1625c85b01a1f2dd9735e9997d0b74d3638c1dc9b14bc0e29139e2f4da8d2"} Oct 04 04:51:09 crc kubenswrapper[4575]: I1004 04:51:09.728079 4575 scope.go:117] "RemoveContainer" containerID="e6a6d9d656e937f48c2e0c38c03877604da45131d454048196bae4e1a3e60a62" Oct 04 04:51:15 crc kubenswrapper[4575]: I1004 04:51:15.384765 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-d6k87" Oct 04 04:51:15 crc kubenswrapper[4575]: I1004 04:51:15.386640 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-d6k87" Oct 04 04:51:15 crc kubenswrapper[4575]: I1004 04:51:15.418742 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-d6k87" Oct 04 04:51:15 crc kubenswrapper[4575]: I1004 04:51:15.788946 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-d6k87" Oct 04 04:51:25 crc kubenswrapper[4575]: I1004 04:51:25.833879 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698"] Oct 04 04:51:25 crc kubenswrapper[4575]: E1004 04:51:25.834624 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3cd2298-f99d-4640-904d-d701c473347b" containerName="registry-server" Oct 04 04:51:25 crc kubenswrapper[4575]: I1004 04:51:25.834648 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3cd2298-f99d-4640-904d-d701c473347b" containerName="registry-server" Oct 04 04:51:25 crc kubenswrapper[4575]: I1004 04:51:25.834771 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3cd2298-f99d-4640-904d-d701c473347b" containerName="registry-server" Oct 04 04:51:25 crc kubenswrapper[4575]: I1004 04:51:25.835533 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:25 crc kubenswrapper[4575]: I1004 04:51:25.837537 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-nwqc7" Oct 04 04:51:25 crc kubenswrapper[4575]: I1004 04:51:25.845744 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698"] Oct 04 04:51:25 crc kubenswrapper[4575]: I1004 04:51:25.977026 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-bundle\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:25 crc kubenswrapper[4575]: I1004 04:51:25.977107 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsbww\" (UniqueName: \"kubernetes.io/projected/15a5cb3d-ea05-4105-9795-7b6839f93323-kube-api-access-vsbww\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:25 crc kubenswrapper[4575]: I1004 04:51:25.977132 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-util\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.078474 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-bundle\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.078560 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsbww\" (UniqueName: \"kubernetes.io/projected/15a5cb3d-ea05-4105-9795-7b6839f93323-kube-api-access-vsbww\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.078601 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-util\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.079136 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-bundle\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.079167 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-util\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.098975 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsbww\" (UniqueName: \"kubernetes.io/projected/15a5cb3d-ea05-4105-9795-7b6839f93323-kube-api-access-vsbww\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.153286 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.592720 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698"] Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.830022 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" event={"ID":"15a5cb3d-ea05-4105-9795-7b6839f93323","Type":"ContainerStarted","Data":"9871cef7127b586cc2a8ae63b1a4f3a713c1e8b75bd8af594bde6d04a09a38d3"} Oct 04 04:51:26 crc kubenswrapper[4575]: I1004 04:51:26.830416 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" event={"ID":"15a5cb3d-ea05-4105-9795-7b6839f93323","Type":"ContainerStarted","Data":"e5e37e1fd9ac4c54a3f620ac8efd9438986d3914b2f2591404f0a9d8c0902158"} Oct 04 04:51:27 crc kubenswrapper[4575]: I1004 04:51:27.837147 4575 generic.go:334] "Generic (PLEG): container finished" podID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerID="9871cef7127b586cc2a8ae63b1a4f3a713c1e8b75bd8af594bde6d04a09a38d3" exitCode=0 Oct 04 04:51:27 crc kubenswrapper[4575]: I1004 04:51:27.837194 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" event={"ID":"15a5cb3d-ea05-4105-9795-7b6839f93323","Type":"ContainerDied","Data":"9871cef7127b586cc2a8ae63b1a4f3a713c1e8b75bd8af594bde6d04a09a38d3"} Oct 04 04:51:28 crc kubenswrapper[4575]: I1004 04:51:28.845455 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" event={"ID":"15a5cb3d-ea05-4105-9795-7b6839f93323","Type":"ContainerStarted","Data":"bb1c7494e9938cc44a8e7a3fdc5080212fe01d8e1cb3f2d9dd3ea0f0067670d2"} Oct 04 04:51:29 crc kubenswrapper[4575]: I1004 04:51:29.854459 4575 generic.go:334] "Generic (PLEG): container finished" podID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerID="bb1c7494e9938cc44a8e7a3fdc5080212fe01d8e1cb3f2d9dd3ea0f0067670d2" exitCode=0 Oct 04 04:51:29 crc kubenswrapper[4575]: I1004 04:51:29.854540 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" event={"ID":"15a5cb3d-ea05-4105-9795-7b6839f93323","Type":"ContainerDied","Data":"bb1c7494e9938cc44a8e7a3fdc5080212fe01d8e1cb3f2d9dd3ea0f0067670d2"} Oct 04 04:51:30 crc kubenswrapper[4575]: I1004 04:51:30.889187 4575 generic.go:334] "Generic (PLEG): container finished" podID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerID="482e7bab7c1303b1391253e3ec413ce35f35c9264643a363887ab3eff9842867" exitCode=0 Oct 04 04:51:30 crc kubenswrapper[4575]: I1004 04:51:30.889262 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" event={"ID":"15a5cb3d-ea05-4105-9795-7b6839f93323","Type":"ContainerDied","Data":"482e7bab7c1303b1391253e3ec413ce35f35c9264643a363887ab3eff9842867"} Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.159320 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.177326 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsbww\" (UniqueName: \"kubernetes.io/projected/15a5cb3d-ea05-4105-9795-7b6839f93323-kube-api-access-vsbww\") pod \"15a5cb3d-ea05-4105-9795-7b6839f93323\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.177663 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-bundle\") pod \"15a5cb3d-ea05-4105-9795-7b6839f93323\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.177701 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-util\") pod \"15a5cb3d-ea05-4105-9795-7b6839f93323\" (UID: \"15a5cb3d-ea05-4105-9795-7b6839f93323\") " Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.178961 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-bundle" (OuterVolumeSpecName: "bundle") pod "15a5cb3d-ea05-4105-9795-7b6839f93323" (UID: "15a5cb3d-ea05-4105-9795-7b6839f93323"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.187767 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a5cb3d-ea05-4105-9795-7b6839f93323-kube-api-access-vsbww" (OuterVolumeSpecName: "kube-api-access-vsbww") pod "15a5cb3d-ea05-4105-9795-7b6839f93323" (UID: "15a5cb3d-ea05-4105-9795-7b6839f93323"). InnerVolumeSpecName "kube-api-access-vsbww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.278413 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsbww\" (UniqueName: \"kubernetes.io/projected/15a5cb3d-ea05-4105-9795-7b6839f93323-kube-api-access-vsbww\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.278472 4575 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.640079 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-util" (OuterVolumeSpecName: "util") pod "15a5cb3d-ea05-4105-9795-7b6839f93323" (UID: "15a5cb3d-ea05-4105-9795-7b6839f93323"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.682490 4575 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15a5cb3d-ea05-4105-9795-7b6839f93323-util\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.902490 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" event={"ID":"15a5cb3d-ea05-4105-9795-7b6839f93323","Type":"ContainerDied","Data":"e5e37e1fd9ac4c54a3f620ac8efd9438986d3914b2f2591404f0a9d8c0902158"} Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.902536 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5e37e1fd9ac4c54a3f620ac8efd9438986d3914b2f2591404f0a9d8c0902158" Oct 04 04:51:32 crc kubenswrapper[4575]: I1004 04:51:32.902539 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.325311 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l"] Oct 04 04:51:38 crc kubenswrapper[4575]: E1004 04:51:38.325645 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerName="extract" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.325662 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerName="extract" Oct 04 04:51:38 crc kubenswrapper[4575]: E1004 04:51:38.325675 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerName="util" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.325685 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerName="util" Oct 04 04:51:38 crc kubenswrapper[4575]: E1004 04:51:38.325710 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerName="pull" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.325719 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerName="pull" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.325868 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="15a5cb3d-ea05-4105-9795-7b6839f93323" containerName="extract" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.326731 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.330879 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-5ns6l" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.358222 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj75r\" (UniqueName: \"kubernetes.io/projected/59e4afb8-7dcb-4c91-b7ff-aac88158d168-kube-api-access-vj75r\") pod \"openstack-operator-controller-operator-6dc5776fd9-l2c9l\" (UID: \"59e4afb8-7dcb-4c91-b7ff-aac88158d168\") " pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.358535 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l"] Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.459858 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj75r\" (UniqueName: \"kubernetes.io/projected/59e4afb8-7dcb-4c91-b7ff-aac88158d168-kube-api-access-vj75r\") pod \"openstack-operator-controller-operator-6dc5776fd9-l2c9l\" (UID: \"59e4afb8-7dcb-4c91-b7ff-aac88158d168\") " pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.491366 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj75r\" (UniqueName: \"kubernetes.io/projected/59e4afb8-7dcb-4c91-b7ff-aac88158d168-kube-api-access-vj75r\") pod \"openstack-operator-controller-operator-6dc5776fd9-l2c9l\" (UID: \"59e4afb8-7dcb-4c91-b7ff-aac88158d168\") " pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.644640 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" Oct 04 04:51:38 crc kubenswrapper[4575]: I1004 04:51:38.957216 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l"] Oct 04 04:51:39 crc kubenswrapper[4575]: I1004 04:51:39.942287 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" event={"ID":"59e4afb8-7dcb-4c91-b7ff-aac88158d168","Type":"ContainerStarted","Data":"64a64ea0bb8fac39b0a92ca19bfdd718776963efb02f3111b9795ab851b0df5a"} Oct 04 04:51:47 crc kubenswrapper[4575]: I1004 04:51:47.017512 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" event={"ID":"59e4afb8-7dcb-4c91-b7ff-aac88158d168","Type":"ContainerStarted","Data":"8017ef800d839f724df804031c8a7746451c9cd3b3d99dcfa2d4731057e2b7d2"} Oct 04 04:51:50 crc kubenswrapper[4575]: I1004 04:51:50.045402 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" event={"ID":"59e4afb8-7dcb-4c91-b7ff-aac88158d168","Type":"ContainerStarted","Data":"cd50e4f5e75b95a7ab64218dbd455a3e8167655ec119ef4912923e8a0d4adef3"} Oct 04 04:51:50 crc kubenswrapper[4575]: I1004 04:51:50.045902 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" Oct 04 04:51:50 crc kubenswrapper[4575]: I1004 04:51:50.081313 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" podStartSLOduration=1.836257952 podStartE2EDuration="12.08129647s" podCreationTimestamp="2025-10-04 04:51:38 +0000 UTC" firstStartedPulling="2025-10-04 04:51:38.976846835 +0000 UTC m=+1050.305405649" lastFinishedPulling="2025-10-04 04:51:49.221885353 +0000 UTC m=+1060.550444167" observedRunningTime="2025-10-04 04:51:50.076124833 +0000 UTC m=+1061.404683647" watchObservedRunningTime="2025-10-04 04:51:50.08129647 +0000 UTC m=+1061.409855284" Oct 04 04:51:51 crc kubenswrapper[4575]: I1004 04:51:51.058001 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-l2c9l" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.433987 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.472244 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.473884 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.476065 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.480746 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-tctq5" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.484251 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qn42c" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.489782 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.493235 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.503879 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.505359 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.511290 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-6269n"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.512684 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.521744 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-ldk8f" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.526838 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-tf8dw" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.527508 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.541901 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-6269n"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.552249 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.553721 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.557868 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-clcng" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.577531 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.595726 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.597051 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.606222 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-576j2" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.611622 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skmbl\" (UniqueName: \"kubernetes.io/projected/51b5294c-5b5b-4689-b5b8-179ecd5cdaf3-kube-api-access-skmbl\") pod \"designate-operator-controller-manager-75dfd9b554-x5sbn\" (UID: \"51b5294c-5b5b-4689-b5b8-179ecd5cdaf3\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.611689 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zxtl\" (UniqueName: \"kubernetes.io/projected/4820b955-13b9-4e2e-a8c6-ec51d1626d36-kube-api-access-8zxtl\") pod \"barbican-operator-controller-manager-5f7c849b98-ngwbj\" (UID: \"4820b955-13b9-4e2e-a8c6-ec51d1626d36\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.611752 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mp9f\" (UniqueName: \"kubernetes.io/projected/dfe68fb9-4ab5-4935-ae7a-805851f52b6c-kube-api-access-5mp9f\") pod \"cinder-operator-controller-manager-7d4d4f8d-fwjkl\" (UID: \"dfe68fb9-4ab5-4935-ae7a-805851f52b6c\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.621190 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.637506 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.639093 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.649403 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.649479 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dzb84" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.668079 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.686677 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.687887 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.692020 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-tp5k5" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.705230 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.712794 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9g8v\" (UniqueName: \"kubernetes.io/projected/295ca73b-c75e-477d-9c47-40c53ef128ea-kube-api-access-p9g8v\") pod \"heat-operator-controller-manager-8f58bc9db-62fpr\" (UID: \"295ca73b-c75e-477d-9c47-40c53ef128ea\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.712847 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m9tl\" (UniqueName: \"kubernetes.io/projected/6854e2d6-7f6c-4bff-9a08-3c91ca5fc895-kube-api-access-7m9tl\") pod \"glance-operator-controller-manager-5568b5d68-6269n\" (UID: \"6854e2d6-7f6c-4bff-9a08-3c91ca5fc895\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.712884 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trw69\" (UniqueName: \"kubernetes.io/projected/696349ff-0270-4925-8d69-c7b5dd5cb77b-kube-api-access-trw69\") pod \"horizon-operator-controller-manager-54876c876f-58mqf\" (UID: \"696349ff-0270-4925-8d69-c7b5dd5cb77b\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.712932 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skmbl\" (UniqueName: \"kubernetes.io/projected/51b5294c-5b5b-4689-b5b8-179ecd5cdaf3-kube-api-access-skmbl\") pod \"designate-operator-controller-manager-75dfd9b554-x5sbn\" (UID: \"51b5294c-5b5b-4689-b5b8-179ecd5cdaf3\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.712963 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zxtl\" (UniqueName: \"kubernetes.io/projected/4820b955-13b9-4e2e-a8c6-ec51d1626d36-kube-api-access-8zxtl\") pod \"barbican-operator-controller-manager-5f7c849b98-ngwbj\" (UID: \"4820b955-13b9-4e2e-a8c6-ec51d1626d36\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.713005 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mp9f\" (UniqueName: \"kubernetes.io/projected/dfe68fb9-4ab5-4935-ae7a-805851f52b6c-kube-api-access-5mp9f\") pod \"cinder-operator-controller-manager-7d4d4f8d-fwjkl\" (UID: \"dfe68fb9-4ab5-4935-ae7a-805851f52b6c\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.714954 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.716394 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.732218 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-mrjkh" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.762847 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mp9f\" (UniqueName: \"kubernetes.io/projected/dfe68fb9-4ab5-4935-ae7a-805851f52b6c-kube-api-access-5mp9f\") pod \"cinder-operator-controller-manager-7d4d4f8d-fwjkl\" (UID: \"dfe68fb9-4ab5-4935-ae7a-805851f52b6c\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.763361 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zxtl\" (UniqueName: \"kubernetes.io/projected/4820b955-13b9-4e2e-a8c6-ec51d1626d36-kube-api-access-8zxtl\") pod \"barbican-operator-controller-manager-5f7c849b98-ngwbj\" (UID: \"4820b955-13b9-4e2e-a8c6-ec51d1626d36\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.768885 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.769918 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.777036 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-4wtcr" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.789689 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.791440 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skmbl\" (UniqueName: \"kubernetes.io/projected/51b5294c-5b5b-4689-b5b8-179ecd5cdaf3-kube-api-access-skmbl\") pod \"designate-operator-controller-manager-75dfd9b554-x5sbn\" (UID: \"51b5294c-5b5b-4689-b5b8-179ecd5cdaf3\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.803555 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.814272 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkqvm\" (UniqueName: \"kubernetes.io/projected/1b4368e7-f41a-43b8-988b-13f5d5388926-kube-api-access-dkqvm\") pod \"ironic-operator-controller-manager-699b87f775-c6lwc\" (UID: \"1b4368e7-f41a-43b8-988b-13f5d5388926\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.814328 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9g8v\" (UniqueName: \"kubernetes.io/projected/295ca73b-c75e-477d-9c47-40c53ef128ea-kube-api-access-p9g8v\") pod \"heat-operator-controller-manager-8f58bc9db-62fpr\" (UID: \"295ca73b-c75e-477d-9c47-40c53ef128ea\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.814369 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m9tl\" (UniqueName: \"kubernetes.io/projected/6854e2d6-7f6c-4bff-9a08-3c91ca5fc895-kube-api-access-7m9tl\") pod \"glance-operator-controller-manager-5568b5d68-6269n\" (UID: \"6854e2d6-7f6c-4bff-9a08-3c91ca5fc895\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.814402 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trw69\" (UniqueName: \"kubernetes.io/projected/696349ff-0270-4925-8d69-c7b5dd5cb77b-kube-api-access-trw69\") pod \"horizon-operator-controller-manager-54876c876f-58mqf\" (UID: \"696349ff-0270-4925-8d69-c7b5dd5cb77b\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.814427 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f920bb1-79e9-42da-88cc-702e70810570-cert\") pod \"infra-operator-controller-manager-658588b8c9-kffhh\" (UID: \"6f920bb1-79e9-42da-88cc-702e70810570\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.814460 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7lhc\" (UniqueName: \"kubernetes.io/projected/6f920bb1-79e9-42da-88cc-702e70810570-kube-api-access-k7lhc\") pod \"infra-operator-controller-manager-658588b8c9-kffhh\" (UID: \"6f920bb1-79e9-42da-88cc-702e70810570\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.815158 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.833040 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.853096 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trw69\" (UniqueName: \"kubernetes.io/projected/696349ff-0270-4925-8d69-c7b5dd5cb77b-kube-api-access-trw69\") pod \"horizon-operator-controller-manager-54876c876f-58mqf\" (UID: \"696349ff-0270-4925-8d69-c7b5dd5cb77b\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.855116 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m9tl\" (UniqueName: \"kubernetes.io/projected/6854e2d6-7f6c-4bff-9a08-3c91ca5fc895-kube-api-access-7m9tl\") pod \"glance-operator-controller-manager-5568b5d68-6269n\" (UID: \"6854e2d6-7f6c-4bff-9a08-3c91ca5fc895\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.859125 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.862508 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.868034 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9g8v\" (UniqueName: \"kubernetes.io/projected/295ca73b-c75e-477d-9c47-40c53ef128ea-kube-api-access-p9g8v\") pod \"heat-operator-controller-manager-8f58bc9db-62fpr\" (UID: \"295ca73b-c75e-477d-9c47-40c53ef128ea\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.873497 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.874423 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.878194 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-j5688" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.895371 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.919871 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ctnf\" (UniqueName: \"kubernetes.io/projected/579db663-7d67-4686-af00-c2bea2e39e98-kube-api-access-6ctnf\") pod \"keystone-operator-controller-manager-655d88ccb9-xhg6m\" (UID: \"579db663-7d67-4686-af00-c2bea2e39e98\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.919922 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f920bb1-79e9-42da-88cc-702e70810570-cert\") pod \"infra-operator-controller-manager-658588b8c9-kffhh\" (UID: \"6f920bb1-79e9-42da-88cc-702e70810570\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.919954 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v7vb\" (UniqueName: \"kubernetes.io/projected/2a08356f-3bae-483a-a361-2b2e4b930ba5-kube-api-access-9v7vb\") pod \"manila-operator-controller-manager-65d89cfd9f-pwfxn\" (UID: \"2a08356f-3bae-483a-a361-2b2e4b930ba5\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.919974 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzkg4\" (UniqueName: \"kubernetes.io/projected/2c106031-0561-4193-ba35-36aec489385a-kube-api-access-xzkg4\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-pqssw\" (UID: \"2c106031-0561-4193-ba35-36aec489385a\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.920000 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7lhc\" (UniqueName: \"kubernetes.io/projected/6f920bb1-79e9-42da-88cc-702e70810570-kube-api-access-k7lhc\") pod \"infra-operator-controller-manager-658588b8c9-kffhh\" (UID: \"6f920bb1-79e9-42da-88cc-702e70810570\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:08 crc kubenswrapper[4575]: E1004 04:52:08.920143 4575 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 04 04:52:08 crc kubenswrapper[4575]: E1004 04:52:08.920247 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f920bb1-79e9-42da-88cc-702e70810570-cert podName:6f920bb1-79e9-42da-88cc-702e70810570 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:09.420213908 +0000 UTC m=+1080.748772722 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6f920bb1-79e9-42da-88cc-702e70810570-cert") pod "infra-operator-controller-manager-658588b8c9-kffhh" (UID: "6f920bb1-79e9-42da-88cc-702e70810570") : secret "infra-operator-webhook-server-cert" not found Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.937956 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkqvm\" (UniqueName: \"kubernetes.io/projected/1b4368e7-f41a-43b8-988b-13f5d5388926-kube-api-access-dkqvm\") pod \"ironic-operator-controller-manager-699b87f775-c6lwc\" (UID: \"1b4368e7-f41a-43b8-988b-13f5d5388926\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.938634 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.964718 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx"] Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.965817 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.979236 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-w58jc" Oct 04 04:52:08 crc kubenswrapper[4575]: I1004 04:52:08.996766 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7lhc\" (UniqueName: \"kubernetes.io/projected/6f920bb1-79e9-42da-88cc-702e70810570-kube-api-access-k7lhc\") pod \"infra-operator-controller-manager-658588b8c9-kffhh\" (UID: \"6f920bb1-79e9-42da-88cc-702e70810570\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.004517 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkqvm\" (UniqueName: \"kubernetes.io/projected/1b4368e7-f41a-43b8-988b-13f5d5388926-kube-api-access-dkqvm\") pod \"ironic-operator-controller-manager-699b87f775-c6lwc\" (UID: \"1b4368e7-f41a-43b8-988b-13f5d5388926\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.008105 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.008409 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.016739 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.018143 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.023402 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.034773 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.047162 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hwck\" (UniqueName: \"kubernetes.io/projected/f5d7fd17-de45-4b0b-9066-0ae5c85e3c67-kube-api-access-9hwck\") pod \"nova-operator-controller-manager-7c7fc454ff-n288p\" (UID: \"f5d7fd17-de45-4b0b-9066-0ae5c85e3c67\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.047203 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55qfl\" (UniqueName: \"kubernetes.io/projected/431117ed-e462-4301-99ca-7a6fef717e89-kube-api-access-55qfl\") pod \"neutron-operator-controller-manager-8d984cc4d-lrjbx\" (UID: \"431117ed-e462-4301-99ca-7a6fef717e89\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.047264 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ctnf\" (UniqueName: \"kubernetes.io/projected/579db663-7d67-4686-af00-c2bea2e39e98-kube-api-access-6ctnf\") pod \"keystone-operator-controller-manager-655d88ccb9-xhg6m\" (UID: \"579db663-7d67-4686-af00-c2bea2e39e98\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.047299 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v7vb\" (UniqueName: \"kubernetes.io/projected/2a08356f-3bae-483a-a361-2b2e4b930ba5-kube-api-access-9v7vb\") pod \"manila-operator-controller-manager-65d89cfd9f-pwfxn\" (UID: \"2a08356f-3bae-483a-a361-2b2e4b930ba5\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.047316 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzkg4\" (UniqueName: \"kubernetes.io/projected/2c106031-0561-4193-ba35-36aec489385a-kube-api-access-xzkg4\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-pqssw\" (UID: \"2c106031-0561-4193-ba35-36aec489385a\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.048285 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-xhd4j" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.080711 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.124933 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ctnf\" (UniqueName: \"kubernetes.io/projected/579db663-7d67-4686-af00-c2bea2e39e98-kube-api-access-6ctnf\") pod \"keystone-operator-controller-manager-655d88ccb9-xhg6m\" (UID: \"579db663-7d67-4686-af00-c2bea2e39e98\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.126163 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzkg4\" (UniqueName: \"kubernetes.io/projected/2c106031-0561-4193-ba35-36aec489385a-kube-api-access-xzkg4\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-pqssw\" (UID: \"2c106031-0561-4193-ba35-36aec489385a\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.129654 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.137567 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v7vb\" (UniqueName: \"kubernetes.io/projected/2a08356f-3bae-483a-a361-2b2e4b930ba5-kube-api-access-9v7vb\") pod \"manila-operator-controller-manager-65d89cfd9f-pwfxn\" (UID: \"2a08356f-3bae-483a-a361-2b2e4b930ba5\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.150523 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.158129 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwknr\" (UniqueName: \"kubernetes.io/projected/2c4bb54c-992e-4341-b480-92aa72f49794-kube-api-access-lwknr\") pod \"ovn-operator-controller-manager-579449c7d5-kkgrw\" (UID: \"2c4bb54c-992e-4341-b480-92aa72f49794\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.158884 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hwck\" (UniqueName: \"kubernetes.io/projected/f5d7fd17-de45-4b0b-9066-0ae5c85e3c67-kube-api-access-9hwck\") pod \"nova-operator-controller-manager-7c7fc454ff-n288p\" (UID: \"f5d7fd17-de45-4b0b-9066-0ae5c85e3c67\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.159786 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55qfl\" (UniqueName: \"kubernetes.io/projected/431117ed-e462-4301-99ca-7a6fef717e89-kube-api-access-55qfl\") pod \"neutron-operator-controller-manager-8d984cc4d-lrjbx\" (UID: \"431117ed-e462-4301-99ca-7a6fef717e89\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.175850 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5br4b" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.222579 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.223054 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.225193 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.240684 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-j7ztm" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.254353 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hwck\" (UniqueName: \"kubernetes.io/projected/f5d7fd17-de45-4b0b-9066-0ae5c85e3c67-kube-api-access-9hwck\") pod \"nova-operator-controller-manager-7c7fc454ff-n288p\" (UID: \"f5d7fd17-de45-4b0b-9066-0ae5c85e3c67\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.260152 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.261397 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.280225 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-qxt6q" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.280433 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.280478 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwknr\" (UniqueName: \"kubernetes.io/projected/2c4bb54c-992e-4341-b480-92aa72f49794-kube-api-access-lwknr\") pod \"ovn-operator-controller-manager-579449c7d5-kkgrw\" (UID: \"2c4bb54c-992e-4341-b480-92aa72f49794\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.280619 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mzrs\" (UniqueName: \"kubernetes.io/projected/6721ebee-1d17-4a50-b450-7b912043621b-kube-api-access-9mzrs\") pod \"octavia-operator-controller-manager-7468f855d8-nnjv5\" (UID: \"6721ebee-1d17-4a50-b450-7b912043621b\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.284395 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.298694 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55qfl\" (UniqueName: \"kubernetes.io/projected/431117ed-e462-4301-99ca-7a6fef717e89-kube-api-access-55qfl\") pod \"neutron-operator-controller-manager-8d984cc4d-lrjbx\" (UID: \"431117ed-e462-4301-99ca-7a6fef717e89\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.305535 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.382649 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-mrjkh" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.383003 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.383419 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b27bf13b-6d49-4f06-ba98-cdcdc0958970-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw\" (UID: \"b27bf13b-6d49-4f06-ba98-cdcdc0958970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.383467 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mzrs\" (UniqueName: \"kubernetes.io/projected/6721ebee-1d17-4a50-b450-7b912043621b-kube-api-access-9mzrs\") pod \"octavia-operator-controller-manager-7468f855d8-nnjv5\" (UID: \"6721ebee-1d17-4a50-b450-7b912043621b\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.383488 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2jqb\" (UniqueName: \"kubernetes.io/projected/b27bf13b-6d49-4f06-ba98-cdcdc0958970-kube-api-access-p2jqb\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw\" (UID: \"b27bf13b-6d49-4f06-ba98-cdcdc0958970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.386060 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.392193 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-w58jc" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.393721 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.406811 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-xhd4j" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.407045 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.408193 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-psw66"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.410002 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.416513 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-qq4gq" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.437315 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-psw66"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.446380 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.447429 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.454331 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-57wbb" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.480929 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mzrs\" (UniqueName: \"kubernetes.io/projected/6721ebee-1d17-4a50-b450-7b912043621b-kube-api-access-9mzrs\") pod \"octavia-operator-controller-manager-7468f855d8-nnjv5\" (UID: \"6721ebee-1d17-4a50-b450-7b912043621b\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.495087 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.498295 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.499623 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f920bb1-79e9-42da-88cc-702e70810570-cert\") pod \"infra-operator-controller-manager-658588b8c9-kffhh\" (UID: \"6f920bb1-79e9-42da-88cc-702e70810570\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.499703 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b27bf13b-6d49-4f06-ba98-cdcdc0958970-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw\" (UID: \"b27bf13b-6d49-4f06-ba98-cdcdc0958970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.499724 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2jqb\" (UniqueName: \"kubernetes.io/projected/b27bf13b-6d49-4f06-ba98-cdcdc0958970-kube-api-access-p2jqb\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw\" (UID: \"b27bf13b-6d49-4f06-ba98-cdcdc0958970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.499780 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rtf8\" (UniqueName: \"kubernetes.io/projected/febb4d47-662e-47e3-ab3b-c78defb6a724-kube-api-access-2rtf8\") pod \"placement-operator-controller-manager-54689d9f88-psw66\" (UID: \"febb4d47-662e-47e3-ab3b-c78defb6a724\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" Oct 04 04:52:09 crc kubenswrapper[4575]: E1004 04:52:09.499980 4575 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 04 04:52:09 crc kubenswrapper[4575]: E1004 04:52:09.500024 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f920bb1-79e9-42da-88cc-702e70810570-cert podName:6f920bb1-79e9-42da-88cc-702e70810570 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:10.500008993 +0000 UTC m=+1081.828567807 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6f920bb1-79e9-42da-88cc-702e70810570-cert") pod "infra-operator-controller-manager-658588b8c9-kffhh" (UID: "6f920bb1-79e9-42da-88cc-702e70810570") : secret "infra-operator-webhook-server-cert" not found Oct 04 04:52:09 crc kubenswrapper[4575]: E1004 04:52:09.501198 4575 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 04:52:09 crc kubenswrapper[4575]: E1004 04:52:09.501276 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b27bf13b-6d49-4f06-ba98-cdcdc0958970-cert podName:b27bf13b-6d49-4f06-ba98-cdcdc0958970 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:10.001256599 +0000 UTC m=+1081.329815413 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b27bf13b-6d49-4f06-ba98-cdcdc0958970-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" (UID: "b27bf13b-6d49-4f06-ba98-cdcdc0958970") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.506512 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-nbj9k" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.543011 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.571264 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.580032 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.584029 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.597401 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-j7ztm" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.599892 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.603071 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2jqb\" (UniqueName: \"kubernetes.io/projected/b27bf13b-6d49-4f06-ba98-cdcdc0958970-kube-api-access-p2jqb\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw\" (UID: \"b27bf13b-6d49-4f06-ba98-cdcdc0958970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.607816 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hf6x\" (UniqueName: \"kubernetes.io/projected/f2f043c4-ba0c-4464-b8f8-fa5fac00f30a-kube-api-access-5hf6x\") pod \"telemetry-operator-controller-manager-5d4d74dd89-kdcqh\" (UID: \"f2f043c4-ba0c-4464-b8f8-fa5fac00f30a\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.607947 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rtf8\" (UniqueName: \"kubernetes.io/projected/febb4d47-662e-47e3-ab3b-c78defb6a724-kube-api-access-2rtf8\") pod \"placement-operator-controller-manager-54689d9f88-psw66\" (UID: \"febb4d47-662e-47e3-ab3b-c78defb6a724\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.608103 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qszps\" (UniqueName: \"kubernetes.io/projected/3e326270-271c-4781-bf2d-cf78cef44891-kube-api-access-qszps\") pod \"swift-operator-controller-manager-6859f9b676-k22tr\" (UID: \"3e326270-271c-4781-bf2d-cf78cef44891\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.623992 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4jxx2" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.664080 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.704353 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rtf8\" (UniqueName: \"kubernetes.io/projected/febb4d47-662e-47e3-ab3b-c78defb6a724-kube-api-access-2rtf8\") pod \"placement-operator-controller-manager-54689d9f88-psw66\" (UID: \"febb4d47-662e-47e3-ab3b-c78defb6a724\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.708357 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwknr\" (UniqueName: \"kubernetes.io/projected/2c4bb54c-992e-4341-b480-92aa72f49794-kube-api-access-lwknr\") pod \"ovn-operator-controller-manager-579449c7d5-kkgrw\" (UID: \"2c4bb54c-992e-4341-b480-92aa72f49794\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.728226 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.738397 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hf6x\" (UniqueName: \"kubernetes.io/projected/f2f043c4-ba0c-4464-b8f8-fa5fac00f30a-kube-api-access-5hf6x\") pod \"telemetry-operator-controller-manager-5d4d74dd89-kdcqh\" (UID: \"f2f043c4-ba0c-4464-b8f8-fa5fac00f30a\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.738533 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clgc9\" (UniqueName: \"kubernetes.io/projected/3841a418-64a3-4065-8d0c-9b1493f1b7b0-kube-api-access-clgc9\") pod \"test-operator-controller-manager-5cd5cb47d7-4j5ln\" (UID: \"3841a418-64a3-4065-8d0c-9b1493f1b7b0\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.738655 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qszps\" (UniqueName: \"kubernetes.io/projected/3e326270-271c-4781-bf2d-cf78cef44891-kube-api-access-qszps\") pod \"swift-operator-controller-manager-6859f9b676-k22tr\" (UID: \"3e326270-271c-4781-bf2d-cf78cef44891\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.738914 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.740988 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.812524 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5br4b" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.813058 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.813812 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-8ckw4" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.839162 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qszps\" (UniqueName: \"kubernetes.io/projected/3e326270-271c-4781-bf2d-cf78cef44891-kube-api-access-qszps\") pod \"swift-operator-controller-manager-6859f9b676-k22tr\" (UID: \"3e326270-271c-4781-bf2d-cf78cef44891\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.856741 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv"] Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.859782 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hf6x\" (UniqueName: \"kubernetes.io/projected/f2f043c4-ba0c-4464-b8f8-fa5fac00f30a-kube-api-access-5hf6x\") pod \"telemetry-operator-controller-manager-5d4d74dd89-kdcqh\" (UID: \"f2f043c4-ba0c-4464-b8f8-fa5fac00f30a\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.860489 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.861283 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clgc9\" (UniqueName: \"kubernetes.io/projected/3841a418-64a3-4065-8d0c-9b1493f1b7b0-kube-api-access-clgc9\") pod \"test-operator-controller-manager-5cd5cb47d7-4j5ln\" (UID: \"3841a418-64a3-4065-8d0c-9b1493f1b7b0\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.861348 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nl45\" (UniqueName: \"kubernetes.io/projected/5a45b0e2-20a2-4d00-a588-0b61682b6bb0-kube-api-access-9nl45\") pod \"watcher-operator-controller-manager-6cbc6dd547-wsbkv\" (UID: \"5a45b0e2-20a2-4d00-a588-0b61682b6bb0\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.928372 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clgc9\" (UniqueName: \"kubernetes.io/projected/3841a418-64a3-4065-8d0c-9b1493f1b7b0-kube-api-access-clgc9\") pod \"test-operator-controller-manager-5cd5cb47d7-4j5ln\" (UID: \"3841a418-64a3-4065-8d0c-9b1493f1b7b0\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.954877 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" Oct 04 04:52:09 crc kubenswrapper[4575]: I1004 04:52:09.962329 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nl45\" (UniqueName: \"kubernetes.io/projected/5a45b0e2-20a2-4d00-a588-0b61682b6bb0-kube-api-access-9nl45\") pod \"watcher-operator-controller-manager-6cbc6dd547-wsbkv\" (UID: \"5a45b0e2-20a2-4d00-a588-0b61682b6bb0\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.049624 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nl45\" (UniqueName: \"kubernetes.io/projected/5a45b0e2-20a2-4d00-a588-0b61682b6bb0-kube-api-access-9nl45\") pod \"watcher-operator-controller-manager-6cbc6dd547-wsbkv\" (UID: \"5a45b0e2-20a2-4d00-a588-0b61682b6bb0\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.068255 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b27bf13b-6d49-4f06-ba98-cdcdc0958970-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw\" (UID: \"b27bf13b-6d49-4f06-ba98-cdcdc0958970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:10 crc kubenswrapper[4575]: E1004 04:52:10.068435 4575 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 04:52:10 crc kubenswrapper[4575]: E1004 04:52:10.068493 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b27bf13b-6d49-4f06-ba98-cdcdc0958970-cert podName:b27bf13b-6d49-4f06-ba98-cdcdc0958970 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:11.068477917 +0000 UTC m=+1082.397036731 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b27bf13b-6d49-4f06-ba98-cdcdc0958970-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" (UID: "b27bf13b-6d49-4f06-ba98-cdcdc0958970") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.110863 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.203518 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs"] Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.215175 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs"] Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.215306 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.230679 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.238173 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5smpd" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.275936 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.330646 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr"] Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.331937 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.337624 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-x72g4" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.351766 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr"] Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.378520 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbca27ef-6c03-4f7b-8682-c30b02a7fcb7-cert\") pod \"openstack-operator-controller-manager-6c6db87b77-wsfrs\" (UID: \"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.378661 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tlxg\" (UniqueName: \"kubernetes.io/projected/dbca27ef-6c03-4f7b-8682-c30b02a7fcb7-kube-api-access-8tlxg\") pod \"openstack-operator-controller-manager-6c6db87b77-wsfrs\" (UID: \"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.515357 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f920bb1-79e9-42da-88cc-702e70810570-cert\") pod \"infra-operator-controller-manager-658588b8c9-kffhh\" (UID: \"6f920bb1-79e9-42da-88cc-702e70810570\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.515442 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tlxg\" (UniqueName: \"kubernetes.io/projected/dbca27ef-6c03-4f7b-8682-c30b02a7fcb7-kube-api-access-8tlxg\") pod \"openstack-operator-controller-manager-6c6db87b77-wsfrs\" (UID: \"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.515518 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9khb\" (UniqueName: \"kubernetes.io/projected/6b2f3b8c-07fa-4a90-81a1-2af4feed6070-kube-api-access-m9khb\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr\" (UID: \"6b2f3b8c-07fa-4a90-81a1-2af4feed6070\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.515596 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbca27ef-6c03-4f7b-8682-c30b02a7fcb7-cert\") pod \"openstack-operator-controller-manager-6c6db87b77-wsfrs\" (UID: \"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:10 crc kubenswrapper[4575]: E1004 04:52:10.515741 4575 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 04 04:52:10 crc kubenswrapper[4575]: E1004 04:52:10.515792 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbca27ef-6c03-4f7b-8682-c30b02a7fcb7-cert podName:dbca27ef-6c03-4f7b-8682-c30b02a7fcb7 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:11.015776689 +0000 UTC m=+1082.344335493 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dbca27ef-6c03-4f7b-8682-c30b02a7fcb7-cert") pod "openstack-operator-controller-manager-6c6db87b77-wsfrs" (UID: "dbca27ef-6c03-4f7b-8682-c30b02a7fcb7") : secret "webhook-server-cert" not found Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.529265 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f920bb1-79e9-42da-88cc-702e70810570-cert\") pod \"infra-operator-controller-manager-658588b8c9-kffhh\" (UID: \"6f920bb1-79e9-42da-88cc-702e70810570\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.607189 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tlxg\" (UniqueName: \"kubernetes.io/projected/dbca27ef-6c03-4f7b-8682-c30b02a7fcb7-kube-api-access-8tlxg\") pod \"openstack-operator-controller-manager-6c6db87b77-wsfrs\" (UID: \"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.617443 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9khb\" (UniqueName: \"kubernetes.io/projected/6b2f3b8c-07fa-4a90-81a1-2af4feed6070-kube-api-access-m9khb\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr\" (UID: \"6b2f3b8c-07fa-4a90-81a1-2af4feed6070\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.643693 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9khb\" (UniqueName: \"kubernetes.io/projected/6b2f3b8c-07fa-4a90-81a1-2af4feed6070-kube-api-access-m9khb\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr\" (UID: \"6b2f3b8c-07fa-4a90-81a1-2af4feed6070\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.684502 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.795511 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dzb84" Oct 04 04:52:10 crc kubenswrapper[4575]: I1004 04:52:10.802776 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.029356 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbca27ef-6c03-4f7b-8682-c30b02a7fcb7-cert\") pod \"openstack-operator-controller-manager-6c6db87b77-wsfrs\" (UID: \"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.055218 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbca27ef-6c03-4f7b-8682-c30b02a7fcb7-cert\") pod \"openstack-operator-controller-manager-6c6db87b77-wsfrs\" (UID: \"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.118050 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.132930 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b27bf13b-6d49-4f06-ba98-cdcdc0958970-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw\" (UID: \"b27bf13b-6d49-4f06-ba98-cdcdc0958970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.137286 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b27bf13b-6d49-4f06-ba98-cdcdc0958970-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw\" (UID: \"b27bf13b-6d49-4f06-ba98-cdcdc0958970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.192701 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.221822 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" event={"ID":"4820b955-13b9-4e2e-a8c6-ec51d1626d36","Type":"ContainerStarted","Data":"2979f3f20402adddb4f2f7665dd409d1e22667213be3a7e1ea86c36c99064170"} Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.274675 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw"] Oct 04 04:52:11 crc kubenswrapper[4575]: W1004 04:52:11.338832 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfe68fb9_4ab5_4935_ae7a_805851f52b6c.slice/crio-37cb06cd3536ed2e031e8674c6c8a7f9e35a49192abe554da6305bb8525588b0 WatchSource:0}: Error finding container 37cb06cd3536ed2e031e8674c6c8a7f9e35a49192abe554da6305bb8525588b0: Status 404 returned error can't find the container with id 37cb06cd3536ed2e031e8674c6c8a7f9e35a49192abe554da6305bb8525588b0 Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.342842 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.385377 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.806030 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.823480 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.858174 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.879072 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-6269n"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.895763 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.906299 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc"] Oct 04 04:52:11 crc kubenswrapper[4575]: W1004 04:52:11.924932 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod696349ff_0270_4925_8d69_c7b5dd5cb77b.slice/crio-971d6d4b4a236c94a57e9f0d448da8ad823e00a468f1a3495665cbb397cbe3a4 WatchSource:0}: Error finding container 971d6d4b4a236c94a57e9f0d448da8ad823e00a468f1a3495665cbb397cbe3a4: Status 404 returned error can't find the container with id 971d6d4b4a236c94a57e9f0d448da8ad823e00a468f1a3495665cbb397cbe3a4 Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.926662 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.938834 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.951440 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln"] Oct 04 04:52:11 crc kubenswrapper[4575]: I1004 04:52:11.988379 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr"] Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.020617 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx"] Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.059971 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh"] Oct 04 04:52:12 crc kubenswrapper[4575]: W1004 04:52:12.129748 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6854e2d6_7f6c_4bff_9a08_3c91ca5fc895.slice/crio-1f60f9e0298df4ce93993cc9a8481aa0f5e568a07a0c01351e27632265ab4f60 WatchSource:0}: Error finding container 1f60f9e0298df4ce93993cc9a8481aa0f5e568a07a0c01351e27632265ab4f60: Status 404 returned error can't find the container with id 1f60f9e0298df4ce93993cc9a8481aa0f5e568a07a0c01351e27632265ab4f60 Oct 04 04:52:12 crc kubenswrapper[4575]: W1004 04:52:12.156174 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod431117ed_e462_4301_99ca_7a6fef717e89.slice/crio-5873d33c075d7647a74b2130ad80b23135ffcd349c3ab345df4f9b7d3af44c6d WatchSource:0}: Error finding container 5873d33c075d7647a74b2130ad80b23135ffcd349c3ab345df4f9b7d3af44c6d: Status 404 returned error can't find the container with id 5873d33c075d7647a74b2130ad80b23135ffcd349c3ab345df4f9b7d3af44c6d Oct 04 04:52:12 crc kubenswrapper[4575]: E1004 04:52:12.214724 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k7lhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-658588b8c9-kffhh_openstack-operators(6f920bb1-79e9-42da-88cc-702e70810570): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.258310 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr"] Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.267355 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh"] Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.294903 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv"] Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.323969 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5"] Oct 04 04:52:12 crc kubenswrapper[4575]: W1004 04:52:12.324832 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b2f3b8c_07fa_4a90_81a1_2af4feed6070.slice/crio-43dc5015274e2147dddcf691fe63af7b91ac3700f8fe3601e4ef899557a57c5c WatchSource:0}: Error finding container 43dc5015274e2147dddcf691fe63af7b91ac3700f8fe3601e4ef899557a57c5c: Status 404 returned error can't find the container with id 43dc5015274e2147dddcf691fe63af7b91ac3700f8fe3601e4ef899557a57c5c Oct 04 04:52:12 crc kubenswrapper[4575]: E1004 04:52:12.325100 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9nl45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6cbc6dd547-wsbkv_openstack-operators(5a45b0e2-20a2-4d00-a588-0b61682b6bb0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:52:12 crc kubenswrapper[4575]: E1004 04:52:12.330784 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m9khb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr_openstack-operators(6b2f3b8c-07fa-4a90-81a1-2af4feed6070): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:52:12 crc kubenswrapper[4575]: E1004 04:52:12.333234 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" podUID="6b2f3b8c-07fa-4a90-81a1-2af4feed6070" Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.349142 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" event={"ID":"2c106031-0561-4193-ba35-36aec489385a","Type":"ContainerStarted","Data":"6fa8541365cacb2629421b8ec94bf5640e92146072863ff493677877c8fa0898"} Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.352369 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw"] Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.353912 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" event={"ID":"295ca73b-c75e-477d-9c47-40c53ef128ea","Type":"ContainerStarted","Data":"1a3a61ba94fa1cb77a620b74bb741cc83c113692df9234a3a91a5a5ae12e5307"} Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.357722 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" event={"ID":"1b4368e7-f41a-43b8-988b-13f5d5388926","Type":"ContainerStarted","Data":"1b121f93e2ad38a6dea0ade74368c2fcba953c71696fd9537f67f485003badb4"} Oct 04 04:52:12 crc kubenswrapper[4575]: W1004 04:52:12.361642 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6721ebee_1d17_4a50_b450_7b912043621b.slice/crio-988590b298cd94a52d9af5dcd74c249c428a9851dff501884abbcd9e7de70cd0 WatchSource:0}: Error finding container 988590b298cd94a52d9af5dcd74c249c428a9851dff501884abbcd9e7de70cd0: Status 404 returned error can't find the container with id 988590b298cd94a52d9af5dcd74c249c428a9851dff501884abbcd9e7de70cd0 Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.361726 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" event={"ID":"2a08356f-3bae-483a-a361-2b2e4b930ba5","Type":"ContainerStarted","Data":"64960c5101865f592ba5600bdace7ef31044c43e4b590fa4306c72cd33e3c5b9"} Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.367171 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" event={"ID":"3841a418-64a3-4065-8d0c-9b1493f1b7b0","Type":"ContainerStarted","Data":"96cc7f3d1f914d361428be2d37d6c15e3c137903e6f977bcafb737350fb42a93"} Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.400132 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-psw66"] Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.407897 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs"] Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.412482 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw"] Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.418005 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" event={"ID":"431117ed-e462-4301-99ca-7a6fef717e89","Type":"ContainerStarted","Data":"5873d33c075d7647a74b2130ad80b23135ffcd349c3ab345df4f9b7d3af44c6d"} Oct 04 04:52:12 crc kubenswrapper[4575]: E1004 04:52:12.422387 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lwknr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-579449c7d5-kkgrw_openstack-operators(2c4bb54c-992e-4341-b480-92aa72f49794): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:52:12 crc kubenswrapper[4575]: E1004 04:52:12.422448 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9mzrs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7468f855d8-nnjv5_openstack-operators(6721ebee-1d17-4a50-b450-7b912043621b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.425673 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" event={"ID":"6854e2d6-7f6c-4bff-9a08-3c91ca5fc895","Type":"ContainerStarted","Data":"1f60f9e0298df4ce93993cc9a8481aa0f5e568a07a0c01351e27632265ab4f60"} Oct 04 04:52:12 crc kubenswrapper[4575]: W1004 04:52:12.430926 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbca27ef_6c03_4f7b_8682_c30b02a7fcb7.slice/crio-7a12ce631a060a949cc07a6b2572e3d3a0fc87efcba614ff56a6927292823420 WatchSource:0}: Error finding container 7a12ce631a060a949cc07a6b2572e3d3a0fc87efcba614ff56a6927292823420: Status 404 returned error can't find the container with id 7a12ce631a060a949cc07a6b2572e3d3a0fc87efcba614ff56a6927292823420 Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.431975 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" event={"ID":"3e326270-271c-4781-bf2d-cf78cef44891","Type":"ContainerStarted","Data":"2f0f5b862014e0d92e71dbe5354003e996ccf6c557880f6d6094d03804c6607f"} Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.484397 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" event={"ID":"6f920bb1-79e9-42da-88cc-702e70810570","Type":"ContainerStarted","Data":"95c489d0abc029c27bcc39de4b0fc40ef10afaa14ad26a31d4a1aaa7336cfb68"} Oct 04 04:52:12 crc kubenswrapper[4575]: E1004 04:52:12.486281 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p2jqb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw_openstack-operators(b27bf13b-6d49-4f06-ba98-cdcdc0958970): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.517715 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" event={"ID":"579db663-7d67-4686-af00-c2bea2e39e98","Type":"ContainerStarted","Data":"34139b2461664d5f7ffde0e75b40522b8128cb42f389c6274b8d9777716f4fa2"} Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.550688 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" event={"ID":"f5d7fd17-de45-4b0b-9066-0ae5c85e3c67","Type":"ContainerStarted","Data":"17171733b874d8830fdfb663dd25058f916a06bca72b0a50882b029cdf8c08ea"} Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.551751 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" event={"ID":"51b5294c-5b5b-4689-b5b8-179ecd5cdaf3","Type":"ContainerStarted","Data":"3ac2f619f3c802fe73ee743c0be4e2f829302a612be27b539fe2c324f159471a"} Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.579024 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" event={"ID":"696349ff-0270-4925-8d69-c7b5dd5cb77b","Type":"ContainerStarted","Data":"971d6d4b4a236c94a57e9f0d448da8ad823e00a468f1a3495665cbb397cbe3a4"} Oct 04 04:52:12 crc kubenswrapper[4575]: I1004 04:52:12.631292 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" event={"ID":"dfe68fb9-4ab5-4935-ae7a-805851f52b6c","Type":"ContainerStarted","Data":"37cb06cd3536ed2e031e8674c6c8a7f9e35a49192abe554da6305bb8525588b0"} Oct 04 04:52:12 crc kubenswrapper[4575]: E1004 04:52:12.986708 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" podUID="6f920bb1-79e9-42da-88cc-702e70810570" Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.155197 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" podUID="5a45b0e2-20a2-4d00-a588-0b61682b6bb0" Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.230049 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" podUID="2c4bb54c-992e-4341-b480-92aa72f49794" Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.230273 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" podUID="6721ebee-1d17-4a50-b450-7b912043621b" Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.299411 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" podUID="b27bf13b-6d49-4f06-ba98-cdcdc0958970" Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.661295 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" event={"ID":"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7","Type":"ContainerStarted","Data":"0574691c9e3bed23c531d8e6c3976ba5b9764a46c017ea43c4642f48824ce2ed"} Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.661344 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" event={"ID":"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7","Type":"ContainerStarted","Data":"633d6bb43aac554f768204f1c2d4513bd25ce6e297bd10e497c03b8f3d2f4e2c"} Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.661357 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" event={"ID":"dbca27ef-6c03-4f7b-8682-c30b02a7fcb7","Type":"ContainerStarted","Data":"7a12ce631a060a949cc07a6b2572e3d3a0fc87efcba614ff56a6927292823420"} Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.662379 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.668389 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" event={"ID":"b27bf13b-6d49-4f06-ba98-cdcdc0958970","Type":"ContainerStarted","Data":"c5abb511905bab58073591d595ad13476bc4cdfbf0764a6d66147a53ae933790"} Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.668499 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" event={"ID":"b27bf13b-6d49-4f06-ba98-cdcdc0958970","Type":"ContainerStarted","Data":"6246247da7e3a3c9fad05b9013257d8c89621ed85d74580ef7ff6e5659a307fc"} Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.669960 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" event={"ID":"febb4d47-662e-47e3-ab3b-c78defb6a724","Type":"ContainerStarted","Data":"cb330ab780a17f8b79fcabb13aadaf5b8413ba541665915bcd1bcff00fafbbca"} Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.670077 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" podUID="b27bf13b-6d49-4f06-ba98-cdcdc0958970" Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.675638 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" event={"ID":"6f920bb1-79e9-42da-88cc-702e70810570","Type":"ContainerStarted","Data":"9d090d27df7519fa96f906ef68ffd7317e9f2cbe5666d9d7b352adec6152bcba"} Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.678315 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" podUID="6f920bb1-79e9-42da-88cc-702e70810570" Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.684451 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" event={"ID":"6721ebee-1d17-4a50-b450-7b912043621b","Type":"ContainerStarted","Data":"2c2cb68e1fe634cbc7dbe7dc75c2efbeedb80055793c95c3fe61151473c5742b"} Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.684508 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" event={"ID":"6721ebee-1d17-4a50-b450-7b912043621b","Type":"ContainerStarted","Data":"988590b298cd94a52d9af5dcd74c249c428a9851dff501884abbcd9e7de70cd0"} Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.686048 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" podUID="6721ebee-1d17-4a50-b450-7b912043621b" Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.687403 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" event={"ID":"6b2f3b8c-07fa-4a90-81a1-2af4feed6070","Type":"ContainerStarted","Data":"43dc5015274e2147dddcf691fe63af7b91ac3700f8fe3601e4ef899557a57c5c"} Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.692855 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" podUID="6b2f3b8c-07fa-4a90-81a1-2af4feed6070" Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.698339 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" event={"ID":"f2f043c4-ba0c-4464-b8f8-fa5fac00f30a","Type":"ContainerStarted","Data":"8baa5eb50aaa6ee9b43687b65e19295426e0bb5db61be5c3bcf8ac59d91fffbf"} Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.711115 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" event={"ID":"2c4bb54c-992e-4341-b480-92aa72f49794","Type":"ContainerStarted","Data":"ad89a144715b9766fac43f43d4da609bb9fa72733be7c4ec1c2bcd66382aeaa5"} Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.711168 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" event={"ID":"2c4bb54c-992e-4341-b480-92aa72f49794","Type":"ContainerStarted","Data":"8a3469568fd5544f0afcc6539d8fca4b3dcf07da9e49ce30f1f56649686703ab"} Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.719472 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" podUID="2c4bb54c-992e-4341-b480-92aa72f49794" Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.721693 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" event={"ID":"5a45b0e2-20a2-4d00-a588-0b61682b6bb0","Type":"ContainerStarted","Data":"b42d50de5bf1878b570a1e22687fc5f93fc70d8fadc0630292ef1a555267ff96"} Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.721741 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" event={"ID":"5a45b0e2-20a2-4d00-a588-0b61682b6bb0","Type":"ContainerStarted","Data":"f100656d7ec84ef720e456659c5a645e6c11a5bcafc8798334ff1f8a920b1efc"} Oct 04 04:52:13 crc kubenswrapper[4575]: E1004 04:52:13.727043 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" podUID="5a45b0e2-20a2-4d00-a588-0b61682b6bb0" Oct 04 04:52:13 crc kubenswrapper[4575]: I1004 04:52:13.729525 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" podStartSLOduration=3.7295046320000003 podStartE2EDuration="3.729504632s" podCreationTimestamp="2025-10-04 04:52:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:52:13.727551407 +0000 UTC m=+1085.056110231" watchObservedRunningTime="2025-10-04 04:52:13.729504632 +0000 UTC m=+1085.058063446" Oct 04 04:52:14 crc kubenswrapper[4575]: E1004 04:52:14.745961 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" podUID="6f920bb1-79e9-42da-88cc-702e70810570" Oct 04 04:52:14 crc kubenswrapper[4575]: E1004 04:52:14.746624 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" podUID="2c4bb54c-992e-4341-b480-92aa72f49794" Oct 04 04:52:14 crc kubenswrapper[4575]: E1004 04:52:14.746679 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" podUID="b27bf13b-6d49-4f06-ba98-cdcdc0958970" Oct 04 04:52:14 crc kubenswrapper[4575]: E1004 04:52:14.746716 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" podUID="5a45b0e2-20a2-4d00-a588-0b61682b6bb0" Oct 04 04:52:14 crc kubenswrapper[4575]: E1004 04:52:14.746727 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" podUID="6b2f3b8c-07fa-4a90-81a1-2af4feed6070" Oct 04 04:52:14 crc kubenswrapper[4575]: E1004 04:52:14.746727 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" podUID="6721ebee-1d17-4a50-b450-7b912043621b" Oct 04 04:52:21 crc kubenswrapper[4575]: I1004 04:52:21.199638 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-wsfrs" Oct 04 04:52:26 crc kubenswrapper[4575]: E1004 04:52:26.705360 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:5f96b563a63494082323bfced089d6589e0c89db43c6a39a2e912c79b1a278fe" Oct 04 04:52:26 crc kubenswrapper[4575]: E1004 04:52:26.705952 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:5f96b563a63494082323bfced089d6589e0c89db43c6a39a2e912c79b1a278fe,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xzkg4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6cd6d7bdf5-pqssw_openstack-operators(2c106031-0561-4193-ba35-36aec489385a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:27 crc kubenswrapper[4575]: E1004 04:52:27.804649 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842" Oct 04 04:52:27 crc kubenswrapper[4575]: E1004 04:52:27.805242 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9hwck,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-7c7fc454ff-n288p_openstack-operators(f5d7fd17-de45-4b0b-9066-0ae5c85e3c67): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:28 crc kubenswrapper[4575]: E1004 04:52:28.274082 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb" Oct 04 04:52:28 crc kubenswrapper[4575]: E1004 04:52:28.274271 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-clgc9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-4j5ln_openstack-operators(3841a418-64a3-4065-8d0c-9b1493f1b7b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:28 crc kubenswrapper[4575]: E1004 04:52:28.962409 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:354a1057bb423082aeda16c0209381a05266e90e30e216522c1462be7d4c4610" Oct 04 04:52:28 crc kubenswrapper[4575]: E1004 04:52:28.962615 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:354a1057bb423082aeda16c0209381a05266e90e30e216522c1462be7d4c4610,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7m9tl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5568b5d68-6269n_openstack-operators(6854e2d6-7f6c-4bff-9a08-3c91ca5fc895): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:29 crc kubenswrapper[4575]: E1004 04:52:29.580865 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed" Oct 04 04:52:29 crc kubenswrapper[4575]: E1004 04:52:29.581064 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qszps,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6859f9b676-k22tr_openstack-operators(3e326270-271c-4781-bf2d-cf78cef44891): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:36 crc kubenswrapper[4575]: E1004 04:52:36.894544 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862" Oct 04 04:52:36 crc kubenswrapper[4575]: E1004 04:52:36.895290 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-55qfl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-8d984cc4d-lrjbx_openstack-operators(431117ed-e462-4301-99ca-7a6fef717e89): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:37 crc kubenswrapper[4575]: E1004 04:52:37.656368 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:445a1332c0eaaa21a5459d3ffe56a8696a6a61131c39dc7bb47571b251a30830" Oct 04 04:52:37 crc kubenswrapper[4575]: E1004 04:52:37.656647 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:445a1332c0eaaa21a5459d3ffe56a8696a6a61131c39dc7bb47571b251a30830,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5mp9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-7d4d4f8d-fwjkl_openstack-operators(dfe68fb9-4ab5-4935-ae7a-805851f52b6c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:38 crc kubenswrapper[4575]: E1004 04:52:38.203357 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e" Oct 04 04:52:38 crc kubenswrapper[4575]: E1004 04:52:38.203580 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5hf6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5d4d74dd89-kdcqh_openstack-operators(f2f043c4-ba0c-4464-b8f8-fa5fac00f30a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:38 crc kubenswrapper[4575]: E1004 04:52:38.709763 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1" Oct 04 04:52:38 crc kubenswrapper[4575]: E1004 04:52:38.710331 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2rtf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-54689d9f88-psw66_openstack-operators(febb4d47-662e-47e3-ab3b-c78defb6a724): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:39 crc kubenswrapper[4575]: E1004 04:52:39.435668 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799" Oct 04 04:52:39 crc kubenswrapper[4575]: E1004 04:52:39.436127 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p2jqb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw_openstack-operators(b27bf13b-6d49-4f06-ba98-cdcdc0958970): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:39 crc kubenswrapper[4575]: E1004 04:52:39.438143 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" podUID="b27bf13b-6d49-4f06-ba98-cdcdc0958970" Oct 04 04:52:40 crc kubenswrapper[4575]: E1004 04:52:40.229088 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:c438734cc669f60ba9d4692fab478cbd326c7de2539d482a21de54a1384ad7ac" Oct 04 04:52:40 crc kubenswrapper[4575]: E1004 04:52:40.229298 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:c438734cc669f60ba9d4692fab478cbd326c7de2539d482a21de54a1384ad7ac,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dkqvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-699b87f775-c6lwc_openstack-operators(1b4368e7-f41a-43b8-988b-13f5d5388926): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:40 crc kubenswrapper[4575]: E1004 04:52:40.776044 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254" Oct 04 04:52:40 crc kubenswrapper[4575]: E1004 04:52:40.776491 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lwknr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-579449c7d5-kkgrw_openstack-operators(2c4bb54c-992e-4341-b480-92aa72f49794): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:52:40 crc kubenswrapper[4575]: E1004 04:52:40.777698 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" podUID="2c4bb54c-992e-4341-b480-92aa72f49794" Oct 04 04:52:42 crc kubenswrapper[4575]: E1004 04:52:42.199433 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" podUID="f5d7fd17-de45-4b0b-9066-0ae5c85e3c67" Oct 04 04:52:42 crc kubenswrapper[4575]: E1004 04:52:42.660567 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" podUID="2c106031-0561-4193-ba35-36aec489385a" Oct 04 04:52:42 crc kubenswrapper[4575]: I1004 04:52:42.986276 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" event={"ID":"431117ed-e462-4301-99ca-7a6fef717e89","Type":"ContainerStarted","Data":"b95b14611352c966d2e18ece5306aa64330b2f1d17b2bd2fc90fe34bcd7a0b8d"} Oct 04 04:52:42 crc kubenswrapper[4575]: I1004 04:52:42.987659 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" event={"ID":"2c106031-0561-4193-ba35-36aec489385a","Type":"ContainerStarted","Data":"a25e62c17df873dd0b360b3dd0c278e8f4f0932e592712aba3fce008cfb5ff95"} Oct 04 04:52:42 crc kubenswrapper[4575]: I1004 04:52:42.995349 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" event={"ID":"6854e2d6-7f6c-4bff-9a08-3c91ca5fc895","Type":"ContainerStarted","Data":"2e282743b4c6d03ad0f571669d7b6954c9524c0460a3554ace9462adf6e94ce8"} Oct 04 04:52:43 crc kubenswrapper[4575]: I1004 04:52:43.003181 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" event={"ID":"febb4d47-662e-47e3-ab3b-c78defb6a724","Type":"ContainerStarted","Data":"80899579af37ad31940a080dd7a4be3055aa2fd9492b79d79c9205e06f5972e9"} Oct 04 04:52:43 crc kubenswrapper[4575]: I1004 04:52:43.011082 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" event={"ID":"3841a418-64a3-4065-8d0c-9b1493f1b7b0","Type":"ContainerStarted","Data":"7301de49e15b07c43bf0ea38d09d9d3e287934d423dcca91c34fd3364a4f7bf6"} Oct 04 04:52:43 crc kubenswrapper[4575]: I1004 04:52:43.021378 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" event={"ID":"f5d7fd17-de45-4b0b-9066-0ae5c85e3c67","Type":"ContainerStarted","Data":"1ea7bc5c244d8887ba817cec1fea9b12cf5394bc197e7ae3a851329d905dda05"} Oct 04 04:52:43 crc kubenswrapper[4575]: E1004 04:52:43.115854 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" podUID="3841a418-64a3-4065-8d0c-9b1493f1b7b0" Oct 04 04:52:43 crc kubenswrapper[4575]: E1004 04:52:43.142362 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" podUID="febb4d47-662e-47e3-ab3b-c78defb6a724" Oct 04 04:52:43 crc kubenswrapper[4575]: E1004 04:52:43.356657 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" podUID="6854e2d6-7f6c-4bff-9a08-3c91ca5fc895" Oct 04 04:52:43 crc kubenswrapper[4575]: E1004 04:52:43.356973 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" podUID="431117ed-e462-4301-99ca-7a6fef717e89" Oct 04 04:52:43 crc kubenswrapper[4575]: E1004 04:52:43.378870 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" podUID="f2f043c4-ba0c-4464-b8f8-fa5fac00f30a" Oct 04 04:52:43 crc kubenswrapper[4575]: E1004 04:52:43.384297 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" podUID="3e326270-271c-4781-bf2d-cf78cef44891" Oct 04 04:52:43 crc kubenswrapper[4575]: E1004 04:52:43.488370 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" podUID="1b4368e7-f41a-43b8-988b-13f5d5388926" Oct 04 04:52:43 crc kubenswrapper[4575]: E1004 04:52:43.845113 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" podUID="dfe68fb9-4ab5-4935-ae7a-805851f52b6c" Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.035912 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" event={"ID":"6f920bb1-79e9-42da-88cc-702e70810570","Type":"ContainerStarted","Data":"a88f32a63b1b4a5129715f010b0b2ac73310e0ac4a08c8fa7a2d7bff5c01d767"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.036189 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.050935 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" event={"ID":"4820b955-13b9-4e2e-a8c6-ec51d1626d36","Type":"ContainerStarted","Data":"e9b481e9cf6ac5dca2091a8f05782ca5d1e79cd986c49f5a10686a31703e1cb4"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.057112 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" event={"ID":"1b4368e7-f41a-43b8-988b-13f5d5388926","Type":"ContainerStarted","Data":"5ffcb89c92b5c5d88992cd69b1369263e58515918b02a1ca5e1936ccb49c73ad"} Oct 04 04:52:44 crc kubenswrapper[4575]: E1004 04:52:44.058707 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:c438734cc669f60ba9d4692fab478cbd326c7de2539d482a21de54a1384ad7ac\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" podUID="1b4368e7-f41a-43b8-988b-13f5d5388926" Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.061522 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" event={"ID":"5a45b0e2-20a2-4d00-a588-0b61682b6bb0","Type":"ContainerStarted","Data":"d50184be372bb352cbd8c5066134ee41cfedd961ed3b14323c4adbbbe849ac6e"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.062135 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.066736 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" event={"ID":"2a08356f-3bae-483a-a361-2b2e4b930ba5","Type":"ContainerStarted","Data":"d836a1cde9ad76ab33a62b33ab492ad6378af4f469d13825352c0c0dfb565ed3"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.068535 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" event={"ID":"6b2f3b8c-07fa-4a90-81a1-2af4feed6070","Type":"ContainerStarted","Data":"53b5f6930102301fde20850bc00f8fc91752a1764a7bd97d5423474bec8a82b4"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.072963 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" event={"ID":"579db663-7d67-4686-af00-c2bea2e39e98","Type":"ContainerStarted","Data":"7190664da8e3fbe6701d1b99e080b79bd762e556fcec84e5c866fd9c6c295e1c"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.075054 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" event={"ID":"f2f043c4-ba0c-4464-b8f8-fa5fac00f30a","Type":"ContainerStarted","Data":"ff7dc80de639595387dae179f320b1a6587b6810ab1134189eb1a7ead1587296"} Oct 04 04:52:44 crc kubenswrapper[4575]: E1004 04:52:44.076355 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" podUID="f2f043c4-ba0c-4464-b8f8-fa5fac00f30a" Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.076477 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" event={"ID":"51b5294c-5b5b-4689-b5b8-179ecd5cdaf3","Type":"ContainerStarted","Data":"ef8672106fd9d567e6defe5e4e680fed268c493260f3af0c0144e6a586eb60be"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.089924 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" event={"ID":"6721ebee-1d17-4a50-b450-7b912043621b","Type":"ContainerStarted","Data":"82e1580330afea658ca9ace529bba26f502d0acedc1270efe4065bedfb428316"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.096239 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" event={"ID":"3e326270-271c-4781-bf2d-cf78cef44891","Type":"ContainerStarted","Data":"5cc0d323207c39db6028be3497602d8c4bce766b1c33fe410ea357c2b4227c8a"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.102092 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" event={"ID":"696349ff-0270-4925-8d69-c7b5dd5cb77b","Type":"ContainerStarted","Data":"0e0772befeb332d45ad9d741afb9b8265df26c6d37739ae7533ca8de7b779668"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.109966 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" event={"ID":"295ca73b-c75e-477d-9c47-40c53ef128ea","Type":"ContainerStarted","Data":"cb2cecae69dbd16e038f6dcac06be39caf9f44e71ad9e2fb154985913951f4d9"} Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.121099 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" event={"ID":"dfe68fb9-4ab5-4935-ae7a-805851f52b6c","Type":"ContainerStarted","Data":"0592d4c25d14134ccd4da02a0382a05ae7359a1dc791950b792ca7d37321113a"} Oct 04 04:52:44 crc kubenswrapper[4575]: E1004 04:52:44.123468 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" podUID="431117ed-e462-4301-99ca-7a6fef717e89" Oct 04 04:52:44 crc kubenswrapper[4575]: E1004 04:52:44.124256 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:445a1332c0eaaa21a5459d3ffe56a8696a6a61131c39dc7bb47571b251a30830\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" podUID="dfe68fb9-4ab5-4935-ae7a-805851f52b6c" Oct 04 04:52:44 crc kubenswrapper[4575]: E1004 04:52:44.125058 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" podUID="febb4d47-662e-47e3-ab3b-c78defb6a724" Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.153187 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" podStartSLOduration=6.098738533 podStartE2EDuration="36.153164359s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.214554315 +0000 UTC m=+1083.543113129" lastFinishedPulling="2025-10-04 04:52:42.268980141 +0000 UTC m=+1113.597538955" observedRunningTime="2025-10-04 04:52:44.132096062 +0000 UTC m=+1115.460654876" watchObservedRunningTime="2025-10-04 04:52:44.153164359 +0000 UTC m=+1115.481723173" Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.327705 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" podStartSLOduration=5.348970043 podStartE2EDuration="35.327675063s" podCreationTimestamp="2025-10-04 04:52:09 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.324955083 +0000 UTC m=+1083.653513897" lastFinishedPulling="2025-10-04 04:52:42.303660103 +0000 UTC m=+1113.632218917" observedRunningTime="2025-10-04 04:52:44.313251044 +0000 UTC m=+1115.641809858" watchObservedRunningTime="2025-10-04 04:52:44.327675063 +0000 UTC m=+1115.656233877" Oct 04 04:52:44 crc kubenswrapper[4575]: I1004 04:52:44.518791 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr" podStartSLOduration=4.719305054 podStartE2EDuration="34.518773407s" podCreationTimestamp="2025-10-04 04:52:10 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.330625503 +0000 UTC m=+1083.659184307" lastFinishedPulling="2025-10-04 04:52:42.130093836 +0000 UTC m=+1113.458652660" observedRunningTime="2025-10-04 04:52:44.514853326 +0000 UTC m=+1115.843412140" watchObservedRunningTime="2025-10-04 04:52:44.518773407 +0000 UTC m=+1115.847332221" Oct 04 04:52:45 crc kubenswrapper[4575]: I1004 04:52:45.129040 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" event={"ID":"51b5294c-5b5b-4689-b5b8-179ecd5cdaf3","Type":"ContainerStarted","Data":"e8b776f412b9e0185d74d514ab09609e3556e2331a68a823ffce327c77287f59"} Oct 04 04:52:45 crc kubenswrapper[4575]: I1004 04:52:45.130224 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" event={"ID":"2a08356f-3bae-483a-a361-2b2e4b930ba5","Type":"ContainerStarted","Data":"3108489ade9b7fe7767147d128becd13f78f2cc37798ff514bcde0df0f1b0f0b"} Oct 04 04:52:45 crc kubenswrapper[4575]: E1004 04:52:45.131745 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:c438734cc669f60ba9d4692fab478cbd326c7de2539d482a21de54a1384ad7ac\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" podUID="1b4368e7-f41a-43b8-988b-13f5d5388926" Oct 04 04:52:45 crc kubenswrapper[4575]: E1004 04:52:45.132291 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:445a1332c0eaaa21a5459d3ffe56a8696a6a61131c39dc7bb47571b251a30830\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" podUID="dfe68fb9-4ab5-4935-ae7a-805851f52b6c" Oct 04 04:52:45 crc kubenswrapper[4575]: E1004 04:52:45.132312 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" podUID="f2f043c4-ba0c-4464-b8f8-fa5fac00f30a" Oct 04 04:52:45 crc kubenswrapper[4575]: I1004 04:52:45.226257 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" podStartSLOduration=7.520691917 podStartE2EDuration="37.226236629s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.4222778 +0000 UTC m=+1083.750836614" lastFinishedPulling="2025-10-04 04:52:42.127822502 +0000 UTC m=+1113.456381326" observedRunningTime="2025-10-04 04:52:45.204345959 +0000 UTC m=+1116.532904773" watchObservedRunningTime="2025-10-04 04:52:45.226236629 +0000 UTC m=+1116.554795443" Oct 04 04:52:46 crc kubenswrapper[4575]: I1004 04:52:46.136852 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" event={"ID":"295ca73b-c75e-477d-9c47-40c53ef128ea","Type":"ContainerStarted","Data":"e73a98485fb780a64aa18354b581c5c5f06f8776d0fd22b83a9d0d9b57967d38"} Oct 04 04:52:46 crc kubenswrapper[4575]: I1004 04:52:46.138455 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" event={"ID":"4820b955-13b9-4e2e-a8c6-ec51d1626d36","Type":"ContainerStarted","Data":"761e49d0ef63585f2fc84d4920a7e3fe5e01e850eb60a43b0350769906425eae"} Oct 04 04:52:47 crc kubenswrapper[4575]: I1004 04:52:47.146793 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" event={"ID":"579db663-7d67-4686-af00-c2bea2e39e98","Type":"ContainerStarted","Data":"d3c78f04b75084108d4b042517c8ba4d1b0dc96c232dccfe26c8464acfae1381"} Oct 04 04:52:47 crc kubenswrapper[4575]: I1004 04:52:47.148856 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" event={"ID":"696349ff-0270-4925-8d69-c7b5dd5cb77b","Type":"ContainerStarted","Data":"eff5f2a9d75c852d809712fca3570aeeef8d43cd80ea4892cb26ad1c6ccf9c6e"} Oct 04 04:52:47 crc kubenswrapper[4575]: I1004 04:52:47.149833 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" Oct 04 04:52:47 crc kubenswrapper[4575]: I1004 04:52:47.149904 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" Oct 04 04:52:47 crc kubenswrapper[4575]: I1004 04:52:47.175088 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" podStartSLOduration=10.470331108 podStartE2EDuration="39.175054848s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.058646218 +0000 UTC m=+1083.387205032" lastFinishedPulling="2025-10-04 04:52:40.763369968 +0000 UTC m=+1112.091928772" observedRunningTime="2025-10-04 04:52:47.170423387 +0000 UTC m=+1118.498982201" watchObservedRunningTime="2025-10-04 04:52:47.175054848 +0000 UTC m=+1118.503613662" Oct 04 04:52:47 crc kubenswrapper[4575]: I1004 04:52:47.189767 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" podStartSLOduration=10.471210212999999 podStartE2EDuration="39.189744664s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.044352513 +0000 UTC m=+1083.372911327" lastFinishedPulling="2025-10-04 04:52:40.762886964 +0000 UTC m=+1112.091445778" observedRunningTime="2025-10-04 04:52:47.186883553 +0000 UTC m=+1118.515442367" watchObservedRunningTime="2025-10-04 04:52:47.189744664 +0000 UTC m=+1118.518303498" Oct 04 04:52:48 crc kubenswrapper[4575]: I1004 04:52:48.157711 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x5sbn" Oct 04 04:52:48 crc kubenswrapper[4575]: I1004 04:52:48.157959 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-pwfxn" Oct 04 04:52:49 crc kubenswrapper[4575]: I1004 04:52:49.184757 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" podStartSLOduration=13.62646239 podStartE2EDuration="41.184739252s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:11.137031789 +0000 UTC m=+1082.465590603" lastFinishedPulling="2025-10-04 04:52:38.695308651 +0000 UTC m=+1110.023867465" observedRunningTime="2025-10-04 04:52:49.178352841 +0000 UTC m=+1120.506911685" watchObservedRunningTime="2025-10-04 04:52:49.184739252 +0000 UTC m=+1120.513298066" Oct 04 04:52:49 crc kubenswrapper[4575]: I1004 04:52:49.603495 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" Oct 04 04:52:49 crc kubenswrapper[4575]: I1004 04:52:49.606417 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-nnjv5" Oct 04 04:52:50 crc kubenswrapper[4575]: I1004 04:52:50.280389 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-wsbkv" Oct 04 04:52:50 crc kubenswrapper[4575]: I1004 04:52:50.811019 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-kffhh" Oct 04 04:52:52 crc kubenswrapper[4575]: I1004 04:52:52.178705 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" Oct 04 04:52:52 crc kubenswrapper[4575]: I1004 04:52:52.179322 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" Oct 04 04:52:52 crc kubenswrapper[4575]: I1004 04:52:52.180746 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" Oct 04 04:52:52 crc kubenswrapper[4575]: I1004 04:52:52.182475 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" Oct 04 04:52:52 crc kubenswrapper[4575]: I1004 04:52:52.198758 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-62fpr" podStartSLOduration=15.486059801 podStartE2EDuration="44.198741057s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.049326934 +0000 UTC m=+1083.377885738" lastFinishedPulling="2025-10-04 04:52:40.76200817 +0000 UTC m=+1112.090566994" observedRunningTime="2025-10-04 04:52:52.195761033 +0000 UTC m=+1123.524319867" watchObservedRunningTime="2025-10-04 04:52:52.198741057 +0000 UTC m=+1123.527299871" Oct 04 04:52:52 crc kubenswrapper[4575]: I1004 04:52:52.210830 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" podStartSLOduration=15.362226533 podStartE2EDuration="44.210813449s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:11.942894559 +0000 UTC m=+1083.271453373" lastFinishedPulling="2025-10-04 04:52:40.791481464 +0000 UTC m=+1112.120040289" observedRunningTime="2025-10-04 04:52:52.20944179 +0000 UTC m=+1123.538000594" watchObservedRunningTime="2025-10-04 04:52:52.210813449 +0000 UTC m=+1123.539372273" Oct 04 04:52:52 crc kubenswrapper[4575]: I1004 04:52:52.250738 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-xhg6m" podStartSLOduration=15.609328324 podStartE2EDuration="44.250715529s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.122624381 +0000 UTC m=+1083.451183195" lastFinishedPulling="2025-10-04 04:52:40.764011586 +0000 UTC m=+1112.092570400" observedRunningTime="2025-10-04 04:52:52.246207492 +0000 UTC m=+1123.574766326" watchObservedRunningTime="2025-10-04 04:52:52.250715529 +0000 UTC m=+1123.579274343" Oct 04 04:52:56 crc kubenswrapper[4575]: E1004 04:52:56.510836 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" podUID="b27bf13b-6d49-4f06-ba98-cdcdc0958970" Oct 04 04:52:56 crc kubenswrapper[4575]: E1004 04:52:56.510966 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" podUID="2c4bb54c-992e-4341-b480-92aa72f49794" Oct 04 04:52:58 crc kubenswrapper[4575]: I1004 04:52:58.805114 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" Oct 04 04:52:58 crc kubenswrapper[4575]: I1004 04:52:58.808467 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-ngwbj" Oct 04 04:52:58 crc kubenswrapper[4575]: I1004 04:52:58.889317 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:52:58 crc kubenswrapper[4575]: I1004 04:52:58.939652 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" Oct 04 04:52:58 crc kubenswrapper[4575]: I1004 04:52:58.941787 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-58mqf" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.240023 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" event={"ID":"3e326270-271c-4781-bf2d-cf78cef44891","Type":"ContainerStarted","Data":"f249f7dd4c1b380fa4a06435b53d6e32d5938abc28f559e642cfa67faabeaf75"} Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.240647 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.255171 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" event={"ID":"f5d7fd17-de45-4b0b-9066-0ae5c85e3c67","Type":"ContainerStarted","Data":"db66e1300fecab6cfb75876b67313bc7ae02f21fea676ba6aeeba2df33a7fcf4"} Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.255332 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.258943 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" event={"ID":"2c106031-0561-4193-ba35-36aec489385a","Type":"ContainerStarted","Data":"d78bb4c8fa98e06420514d181674724c05b5f9503080c8ef7fce6363c096a10a"} Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.259046 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.262401 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" event={"ID":"6854e2d6-7f6c-4bff-9a08-3c91ca5fc895","Type":"ContainerStarted","Data":"21f5c5043631972b94e44014e55dbcc3027d6fc513e84a8ffc56006b94741b3c"} Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.262626 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.264897 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" event={"ID":"3841a418-64a3-4065-8d0c-9b1493f1b7b0","Type":"ContainerStarted","Data":"df02fb02bc739a70d993cc426332095d2b7cd5c09e47640bf16eb05fca3e0ae4"} Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.265349 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.266412 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" podStartSLOduration=4.653367835 podStartE2EDuration="52.266388647s" podCreationTimestamp="2025-10-04 04:52:09 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.048955364 +0000 UTC m=+1083.377514188" lastFinishedPulling="2025-10-04 04:52:59.661976196 +0000 UTC m=+1130.990535000" observedRunningTime="2025-10-04 04:53:01.263212777 +0000 UTC m=+1132.591771591" watchObservedRunningTime="2025-10-04 04:53:01.266388647 +0000 UTC m=+1132.594947461" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.286684 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" podStartSLOduration=5.495045391 podStartE2EDuration="53.286663102s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:11.869399917 +0000 UTC m=+1083.197958731" lastFinishedPulling="2025-10-04 04:52:59.661017628 +0000 UTC m=+1130.989576442" observedRunningTime="2025-10-04 04:53:01.282067691 +0000 UTC m=+1132.610626515" watchObservedRunningTime="2025-10-04 04:53:01.286663102 +0000 UTC m=+1132.615221926" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.313930 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" podStartSLOduration=4.952647984 podStartE2EDuration="53.313912363s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:11.304801032 +0000 UTC m=+1082.633359846" lastFinishedPulling="2025-10-04 04:52:59.666065411 +0000 UTC m=+1130.994624225" observedRunningTime="2025-10-04 04:53:01.297816928 +0000 UTC m=+1132.626375752" watchObservedRunningTime="2025-10-04 04:53:01.313912363 +0000 UTC m=+1132.642471177" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.333099 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" podStartSLOduration=5.804708812 podStartE2EDuration="53.333072966s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.133336714 +0000 UTC m=+1083.461895528" lastFinishedPulling="2025-10-04 04:52:59.661700868 +0000 UTC m=+1130.990259682" observedRunningTime="2025-10-04 04:53:01.316207259 +0000 UTC m=+1132.644766073" watchObservedRunningTime="2025-10-04 04:53:01.333072966 +0000 UTC m=+1132.661631780" Oct 04 04:53:01 crc kubenswrapper[4575]: I1004 04:53:01.333525 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" podStartSLOduration=4.803357854 podStartE2EDuration="52.333518959s" podCreationTimestamp="2025-10-04 04:52:09 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.130940216 +0000 UTC m=+1083.459499030" lastFinishedPulling="2025-10-04 04:52:59.661101321 +0000 UTC m=+1130.989660135" observedRunningTime="2025-10-04 04:53:01.330756991 +0000 UTC m=+1132.659315805" watchObservedRunningTime="2025-10-04 04:53:01.333518959 +0000 UTC m=+1132.662077793" Oct 04 04:53:03 crc kubenswrapper[4575]: I1004 04:53:03.280931 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" event={"ID":"1b4368e7-f41a-43b8-988b-13f5d5388926","Type":"ContainerStarted","Data":"fa252cce067995ecaf76091e46a4abc6833a7ea34d4375227f9ce1ce79bcffb3"} Oct 04 04:53:03 crc kubenswrapper[4575]: I1004 04:53:03.281667 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" Oct 04 04:53:03 crc kubenswrapper[4575]: I1004 04:53:03.283272 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" event={"ID":"431117ed-e462-4301-99ca-7a6fef717e89","Type":"ContainerStarted","Data":"71813ed476b2d494d3ca706cec5195f567eecf0b9a23bbdaf0d7f08caece5f7e"} Oct 04 04:53:03 crc kubenswrapper[4575]: I1004 04:53:03.283547 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" Oct 04 04:53:03 crc kubenswrapper[4575]: I1004 04:53:03.285246 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" event={"ID":"febb4d47-662e-47e3-ab3b-c78defb6a724","Type":"ContainerStarted","Data":"d71903772b22530836a476c12de0fac468bdd8a9490f1144a1867edb36399a24"} Oct 04 04:53:03 crc kubenswrapper[4575]: I1004 04:53:03.285763 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" Oct 04 04:53:03 crc kubenswrapper[4575]: I1004 04:53:03.354752 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" podStartSLOduration=4.800799394 podStartE2EDuration="55.35472958s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.044825197 +0000 UTC m=+1083.373384011" lastFinishedPulling="2025-10-04 04:53:02.598755383 +0000 UTC m=+1133.927314197" observedRunningTime="2025-10-04 04:53:03.326222142 +0000 UTC m=+1134.654780966" watchObservedRunningTime="2025-10-04 04:53:03.35472958 +0000 UTC m=+1134.683288404" Oct 04 04:53:03 crc kubenswrapper[4575]: I1004 04:53:03.369371 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" podStartSLOduration=4.262894434 podStartE2EDuration="54.369344384s" podCreationTimestamp="2025-10-04 04:52:09 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.485643715 +0000 UTC m=+1083.814202539" lastFinishedPulling="2025-10-04 04:53:02.592093675 +0000 UTC m=+1133.920652489" observedRunningTime="2025-10-04 04:53:03.35297347 +0000 UTC m=+1134.681532284" watchObservedRunningTime="2025-10-04 04:53:03.369344384 +0000 UTC m=+1134.697903198" Oct 04 04:53:03 crc kubenswrapper[4575]: I1004 04:53:03.373850 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" podStartSLOduration=4.985766442 podStartE2EDuration="55.373842531s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.205076656 +0000 UTC m=+1083.533635470" lastFinishedPulling="2025-10-04 04:53:02.593152745 +0000 UTC m=+1133.921711559" observedRunningTime="2025-10-04 04:53:03.368194281 +0000 UTC m=+1134.696753105" watchObservedRunningTime="2025-10-04 04:53:03.373842531 +0000 UTC m=+1134.702401345" Oct 04 04:53:04 crc kubenswrapper[4575]: I1004 04:53:04.293363 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" event={"ID":"dfe68fb9-4ab5-4935-ae7a-805851f52b6c","Type":"ContainerStarted","Data":"88b8cc7146a988f0fbed59f378325049b0e082f98bb27c230181b51d9e7ca9e5"} Oct 04 04:53:04 crc kubenswrapper[4575]: I1004 04:53:04.293632 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" Oct 04 04:53:04 crc kubenswrapper[4575]: I1004 04:53:04.294996 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" event={"ID":"f2f043c4-ba0c-4464-b8f8-fa5fac00f30a","Type":"ContainerStarted","Data":"94b1ed8c45ac854b3e8a96a85e80c68aa81a7d0261c645a7ed9de8a929df3a8d"} Oct 04 04:53:04 crc kubenswrapper[4575]: I1004 04:53:04.320062 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" podStartSLOduration=3.880872722 podStartE2EDuration="56.320045636s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:11.347828201 +0000 UTC m=+1082.676387015" lastFinishedPulling="2025-10-04 04:53:03.787001115 +0000 UTC m=+1135.115559929" observedRunningTime="2025-10-04 04:53:04.315107966 +0000 UTC m=+1135.643666780" watchObservedRunningTime="2025-10-04 04:53:04.320045636 +0000 UTC m=+1135.648604450" Oct 04 04:53:04 crc kubenswrapper[4575]: I1004 04:53:04.335152 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" podStartSLOduration=3.873400531 podStartE2EDuration="55.335134414s" podCreationTimestamp="2025-10-04 04:52:09 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.322911595 +0000 UTC m=+1083.651470409" lastFinishedPulling="2025-10-04 04:53:03.784645478 +0000 UTC m=+1135.113204292" observedRunningTime="2025-10-04 04:53:04.334948898 +0000 UTC m=+1135.663507722" watchObservedRunningTime="2025-10-04 04:53:04.335134414 +0000 UTC m=+1135.663693238" Oct 04 04:53:08 crc kubenswrapper[4575]: I1004 04:53:08.878483 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-6269n" Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.013796 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-c6lwc" Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.226901 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-pqssw" Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.357849 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" event={"ID":"2c4bb54c-992e-4341-b480-92aa72f49794","Type":"ContainerStarted","Data":"2452d6a860db24cc20b3e4e05c2ca7063eba6f6845dddec5f61db5c8a02fb416"} Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.358951 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.395759 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lrjbx" Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.402323 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" podStartSLOduration=4.931253728 podStartE2EDuration="1m1.402297033s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.422228778 +0000 UTC m=+1083.750787592" lastFinishedPulling="2025-10-04 04:53:08.893272083 +0000 UTC m=+1140.221830897" observedRunningTime="2025-10-04 04:53:09.399471133 +0000 UTC m=+1140.728029957" watchObservedRunningTime="2025-10-04 04:53:09.402297033 +0000 UTC m=+1140.730855847" Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.415568 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-n288p" Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.733819 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-psw66" Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.865027 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-k22tr" Oct 04 04:53:09 crc kubenswrapper[4575]: I1004 04:53:09.971931 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4j5ln" Oct 04 04:53:10 crc kubenswrapper[4575]: I1004 04:53:10.123447 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" Oct 04 04:53:10 crc kubenswrapper[4575]: I1004 04:53:10.137363 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-kdcqh" Oct 04 04:53:10 crc kubenswrapper[4575]: I1004 04:53:10.367458 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" event={"ID":"b27bf13b-6d49-4f06-ba98-cdcdc0958970","Type":"ContainerStarted","Data":"1758fe79db5c59e181e42ac1a81ff8be42e6add925d30abdb219516e0a4df466"} Oct 04 04:53:10 crc kubenswrapper[4575]: I1004 04:53:10.396794 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" podStartSLOduration=4.870907159 podStartE2EDuration="1m2.396740286s" podCreationTimestamp="2025-10-04 04:52:08 +0000 UTC" firstStartedPulling="2025-10-04 04:52:12.485912502 +0000 UTC m=+1083.814471316" lastFinishedPulling="2025-10-04 04:53:10.011745629 +0000 UTC m=+1141.340304443" observedRunningTime="2025-10-04 04:53:10.396252952 +0000 UTC m=+1141.724811786" watchObservedRunningTime="2025-10-04 04:53:10.396740286 +0000 UTC m=+1141.725299100" Oct 04 04:53:11 crc kubenswrapper[4575]: I1004 04:53:11.386079 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:53:18 crc kubenswrapper[4575]: I1004 04:53:18.836710 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-fwjkl" Oct 04 04:53:19 crc kubenswrapper[4575]: I1004 04:53:19.818654 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-kkgrw" Oct 04 04:53:21 crc kubenswrapper[4575]: I1004 04:53:21.392243 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw" Oct 04 04:53:38 crc kubenswrapper[4575]: I1004 04:53:38.446741 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:53:38 crc kubenswrapper[4575]: I1004 04:53:38.447403 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.306626 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4wmw2"] Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.307881 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.312066 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.312860 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-5z9qf" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.317537 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.317716 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.371075 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4wmw2"] Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.441186 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mh9hm"] Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.443952 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.446798 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.453125 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mh9hm"] Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.457551 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-config\") pod \"dnsmasq-dns-78dd6ddcc-mh9hm\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.457622 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fnrn\" (UniqueName: \"kubernetes.io/projected/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-kube-api-access-6fnrn\") pod \"dnsmasq-dns-675f4bcbfc-4wmw2\" (UID: \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.457644 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-config\") pod \"dnsmasq-dns-675f4bcbfc-4wmw2\" (UID: \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.457784 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mh9hm\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.457940 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbgh4\" (UniqueName: \"kubernetes.io/projected/d2885fac-69b9-4c11-a52c-d88188cee016-kube-api-access-xbgh4\") pod \"dnsmasq-dns-78dd6ddcc-mh9hm\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.559859 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fnrn\" (UniqueName: \"kubernetes.io/projected/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-kube-api-access-6fnrn\") pod \"dnsmasq-dns-675f4bcbfc-4wmw2\" (UID: \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.560476 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-config\") pod \"dnsmasq-dns-675f4bcbfc-4wmw2\" (UID: \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.560520 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mh9hm\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.560758 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbgh4\" (UniqueName: \"kubernetes.io/projected/d2885fac-69b9-4c11-a52c-d88188cee016-kube-api-access-xbgh4\") pod \"dnsmasq-dns-78dd6ddcc-mh9hm\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.560813 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-config\") pod \"dnsmasq-dns-78dd6ddcc-mh9hm\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.562149 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mh9hm\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.562964 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-config\") pod \"dnsmasq-dns-675f4bcbfc-4wmw2\" (UID: \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.564336 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-config\") pod \"dnsmasq-dns-78dd6ddcc-mh9hm\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.591881 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fnrn\" (UniqueName: \"kubernetes.io/projected/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-kube-api-access-6fnrn\") pod \"dnsmasq-dns-675f4bcbfc-4wmw2\" (UID: \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.603725 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbgh4\" (UniqueName: \"kubernetes.io/projected/d2885fac-69b9-4c11-a52c-d88188cee016-kube-api-access-xbgh4\") pod \"dnsmasq-dns-78dd6ddcc-mh9hm\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.627954 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:53:39 crc kubenswrapper[4575]: I1004 04:53:39.769966 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:53:40 crc kubenswrapper[4575]: I1004 04:53:40.095359 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4wmw2"] Oct 04 04:53:40 crc kubenswrapper[4575]: I1004 04:53:40.308276 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mh9hm"] Oct 04 04:53:40 crc kubenswrapper[4575]: W1004 04:53:40.315793 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2885fac_69b9_4c11_a52c_d88188cee016.slice/crio-1b1ca972adfddf6799e4cc6d9a678ee97ae151b32957f3fd2211b89bbaa296b2 WatchSource:0}: Error finding container 1b1ca972adfddf6799e4cc6d9a678ee97ae151b32957f3fd2211b89bbaa296b2: Status 404 returned error can't find the container with id 1b1ca972adfddf6799e4cc6d9a678ee97ae151b32957f3fd2211b89bbaa296b2 Oct 04 04:53:40 crc kubenswrapper[4575]: I1004 04:53:40.572945 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" event={"ID":"d2885fac-69b9-4c11-a52c-d88188cee016","Type":"ContainerStarted","Data":"1b1ca972adfddf6799e4cc6d9a678ee97ae151b32957f3fd2211b89bbaa296b2"} Oct 04 04:53:40 crc kubenswrapper[4575]: I1004 04:53:40.575900 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" event={"ID":"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1","Type":"ContainerStarted","Data":"ecf3315f5d2e9510f41d93dd55499ebe0fba989470c7657c59b0405535d419cd"} Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.550204 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4wmw2"] Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.584762 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6cgqb"] Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.586347 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.615247 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6cgqb"] Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.711424 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-6cgqb\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.711517 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46zzx\" (UniqueName: \"kubernetes.io/projected/5569b565-1561-4c80-84b4-fa48965c1d1f-kube-api-access-46zzx\") pod \"dnsmasq-dns-666b6646f7-6cgqb\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.711548 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-config\") pod \"dnsmasq-dns-666b6646f7-6cgqb\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.812807 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-6cgqb\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.812910 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46zzx\" (UniqueName: \"kubernetes.io/projected/5569b565-1561-4c80-84b4-fa48965c1d1f-kube-api-access-46zzx\") pod \"dnsmasq-dns-666b6646f7-6cgqb\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.812956 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-config\") pod \"dnsmasq-dns-666b6646f7-6cgqb\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.814127 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-config\") pod \"dnsmasq-dns-666b6646f7-6cgqb\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.814555 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-6cgqb\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.867857 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46zzx\" (UniqueName: \"kubernetes.io/projected/5569b565-1561-4c80-84b4-fa48965c1d1f-kube-api-access-46zzx\") pod \"dnsmasq-dns-666b6646f7-6cgqb\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.916367 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:53:42 crc kubenswrapper[4575]: I1004 04:53:42.961820 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mh9hm"] Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.002258 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-48z8b"] Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.006135 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.032189 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-48z8b"] Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.128252 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-48z8b\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.128311 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-config\") pod \"dnsmasq-dns-57d769cc4f-48z8b\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.128390 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px5rn\" (UniqueName: \"kubernetes.io/projected/4c489f23-9a31-437f-ae9c-3f421ff5aace-kube-api-access-px5rn\") pod \"dnsmasq-dns-57d769cc4f-48z8b\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.230432 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px5rn\" (UniqueName: \"kubernetes.io/projected/4c489f23-9a31-437f-ae9c-3f421ff5aace-kube-api-access-px5rn\") pod \"dnsmasq-dns-57d769cc4f-48z8b\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.230507 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-48z8b\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.230550 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-config\") pod \"dnsmasq-dns-57d769cc4f-48z8b\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.231894 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-config\") pod \"dnsmasq-dns-57d769cc4f-48z8b\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.232144 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-48z8b\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.264681 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px5rn\" (UniqueName: \"kubernetes.io/projected/4c489f23-9a31-437f-ae9c-3f421ff5aace-kube-api-access-px5rn\") pod \"dnsmasq-dns-57d769cc4f-48z8b\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.364212 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.706041 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6cgqb"] Oct 04 04:53:43 crc kubenswrapper[4575]: W1004 04:53:43.718456 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5569b565_1561_4c80_84b4_fa48965c1d1f.slice/crio-3fc8215aee7adf6d541e3f877b9ae090562b7dc568937edb620e775339a3512b WatchSource:0}: Error finding container 3fc8215aee7adf6d541e3f877b9ae090562b7dc568937edb620e775339a3512b: Status 404 returned error can't find the container with id 3fc8215aee7adf6d541e3f877b9ae090562b7dc568937edb620e775339a3512b Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.805859 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.807326 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.810641 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-p9gkh" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.811006 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.811279 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.811378 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.811417 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.811611 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.816639 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.839254 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.913082 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-48z8b"] Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.952078 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.952420 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.952462 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.952493 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-config-data\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.952522 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c16a4e4c-f361-42ef-9af1-e7601b14ca69-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.952550 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c16a4e4c-f361-42ef-9af1-e7601b14ca69-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.952830 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.953160 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.953258 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.953301 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:43 crc kubenswrapper[4575]: I1004 04:53:43.953339 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28d8b\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-kube-api-access-28d8b\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054216 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c16a4e4c-f361-42ef-9af1-e7601b14ca69-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054262 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054286 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054303 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054358 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054386 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28d8b\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-kube-api-access-28d8b\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054411 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054436 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054462 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054480 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-config-data\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054502 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c16a4e4c-f361-42ef-9af1-e7601b14ca69-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.054823 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.055292 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.055631 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.059191 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.060314 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-config-data\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.060421 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.064466 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c16a4e4c-f361-42ef-9af1-e7601b14ca69-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.064504 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c16a4e4c-f361-42ef-9af1-e7601b14ca69-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.065141 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.080856 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.089954 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.108202 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28d8b\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-kube-api-access-28d8b\") pod \"rabbitmq-server-0\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.129430 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.144959 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.152205 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.158261 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zks2v" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.159115 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.159362 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.159665 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.159853 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.160229 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.162517 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.193106 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.264874 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.264933 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.264962 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.264993 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.265010 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.265052 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.265075 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.265097 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jksjm\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-kube-api-access-jksjm\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.265133 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.265162 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.265183 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.366758 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.366974 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.367208 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.367236 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.367309 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.367335 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.367361 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jksjm\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-kube-api-access-jksjm\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.367406 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.367431 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.367449 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.367528 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.368210 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.368476 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.372044 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.372076 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.372517 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.373115 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.373199 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.373543 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.382603 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.384366 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.387684 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jksjm\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-kube-api-access-jksjm\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.396684 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.489410 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.662394 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" event={"ID":"5569b565-1561-4c80-84b4-fa48965c1d1f","Type":"ContainerStarted","Data":"3fc8215aee7adf6d541e3f877b9ae090562b7dc568937edb620e775339a3512b"} Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.663780 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" event={"ID":"4c489f23-9a31-437f-ae9c-3f421ff5aace","Type":"ContainerStarted","Data":"6cfd0caba94e91ca91a3c44c8e72fe619050e7541bfe4179f32b63d2ca6907b2"} Oct 04 04:53:44 crc kubenswrapper[4575]: I1004 04:53:44.673508 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:53:45 crc kubenswrapper[4575]: I1004 04:53:45.063454 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:53:45 crc kubenswrapper[4575]: I1004 04:53:45.672297 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c16a4e4c-f361-42ef-9af1-e7601b14ca69","Type":"ContainerStarted","Data":"76184077f91faaed087ee31a80894f297f5cd645cb59f18b89e775f08eb82455"} Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.225574 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.226811 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.228659 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.229925 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-h452g" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.230088 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.230195 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.233969 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.248544 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.253396 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.402056 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/21199967-fd4f-43af-acd4-653828f2335a-kolla-config\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.402128 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ctt\" (UniqueName: \"kubernetes.io/projected/21199967-fd4f-43af-acd4-653828f2335a-kube-api-access-j9ctt\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.402152 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/21199967-fd4f-43af-acd4-653828f2335a-config-data-default\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.402197 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/21199967-fd4f-43af-acd4-653828f2335a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.402216 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.402261 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21199967-fd4f-43af-acd4-653828f2335a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.402286 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/21199967-fd4f-43af-acd4-653828f2335a-secrets\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.402379 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21199967-fd4f-43af-acd4-653828f2335a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.402529 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/21199967-fd4f-43af-acd4-653828f2335a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.504070 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ctt\" (UniqueName: \"kubernetes.io/projected/21199967-fd4f-43af-acd4-653828f2335a-kube-api-access-j9ctt\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.504136 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/21199967-fd4f-43af-acd4-653828f2335a-config-data-default\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.504195 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/21199967-fd4f-43af-acd4-653828f2335a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.504231 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.504256 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21199967-fd4f-43af-acd4-653828f2335a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.504307 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/21199967-fd4f-43af-acd4-653828f2335a-secrets\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.504345 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21199967-fd4f-43af-acd4-653828f2335a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.504396 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/21199967-fd4f-43af-acd4-653828f2335a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.504441 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/21199967-fd4f-43af-acd4-653828f2335a-kolla-config\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.505476 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/21199967-fd4f-43af-acd4-653828f2335a-kolla-config\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.506091 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/21199967-fd4f-43af-acd4-653828f2335a-config-data-default\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.506499 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/21199967-fd4f-43af-acd4-653828f2335a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.506698 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.506793 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21199967-fd4f-43af-acd4-653828f2335a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.510931 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/21199967-fd4f-43af-acd4-653828f2335a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.511200 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21199967-fd4f-43af-acd4-653828f2335a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.519877 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/21199967-fd4f-43af-acd4-653828f2335a-secrets\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.526206 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ctt\" (UniqueName: \"kubernetes.io/projected/21199967-fd4f-43af-acd4-653828f2335a-kube-api-access-j9ctt\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.541856 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"21199967-fd4f-43af-acd4-653828f2335a\") " pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.553744 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.648830 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.650548 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.656243 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.656548 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.656939 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-6n7f5" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.656964 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.683120 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.812394 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04440b88-f2b8-418c-9676-714f0e8b7112-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.812931 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/04440b88-f2b8-418c-9676-714f0e8b7112-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.812985 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04440b88-f2b8-418c-9676-714f0e8b7112-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.813019 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.813104 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh9f5\" (UniqueName: \"kubernetes.io/projected/04440b88-f2b8-418c-9676-714f0e8b7112-kube-api-access-kh9f5\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.813142 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/04440b88-f2b8-418c-9676-714f0e8b7112-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.813180 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/04440b88-f2b8-418c-9676-714f0e8b7112-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.813228 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/04440b88-f2b8-418c-9676-714f0e8b7112-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.813303 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/04440b88-f2b8-418c-9676-714f0e8b7112-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.914909 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04440b88-f2b8-418c-9676-714f0e8b7112-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.914999 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/04440b88-f2b8-418c-9676-714f0e8b7112-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.915041 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04440b88-f2b8-418c-9676-714f0e8b7112-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.915069 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.915100 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh9f5\" (UniqueName: \"kubernetes.io/projected/04440b88-f2b8-418c-9676-714f0e8b7112-kube-api-access-kh9f5\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.915128 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/04440b88-f2b8-418c-9676-714f0e8b7112-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.915157 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/04440b88-f2b8-418c-9676-714f0e8b7112-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.915199 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/04440b88-f2b8-418c-9676-714f0e8b7112-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.915230 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/04440b88-f2b8-418c-9676-714f0e8b7112-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.915653 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.916896 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/04440b88-f2b8-418c-9676-714f0e8b7112-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.917264 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04440b88-f2b8-418c-9676-714f0e8b7112-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.917855 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/04440b88-f2b8-418c-9676-714f0e8b7112-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.917982 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/04440b88-f2b8-418c-9676-714f0e8b7112-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.922715 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04440b88-f2b8-418c-9676-714f0e8b7112-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.922936 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/04440b88-f2b8-418c-9676-714f0e8b7112-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.945421 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/04440b88-f2b8-418c-9676-714f0e8b7112-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.965072 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.965792 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh9f5\" (UniqueName: \"kubernetes.io/projected/04440b88-f2b8-418c-9676-714f0e8b7112-kube-api-access-kh9f5\") pod \"openstack-cell1-galera-0\" (UID: \"04440b88-f2b8-418c-9676-714f0e8b7112\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.991031 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 04 04:53:46 crc kubenswrapper[4575]: I1004 04:53:46.999809 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.007034 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-wqgsk" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.007299 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.011028 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.013177 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.118206 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-kolla-config\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.118268 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.118302 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-config-data\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.118387 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcmbg\" (UniqueName: \"kubernetes.io/projected/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-kube-api-access-bcmbg\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.118692 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.220135 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.220246 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-kolla-config\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.220271 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.220397 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-config-data\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.221989 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-config-data\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.222319 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-kolla-config\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.222366 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcmbg\" (UniqueName: \"kubernetes.io/projected/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-kube-api-access-bcmbg\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.223756 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.226200 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.244561 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcmbg\" (UniqueName: \"kubernetes.io/projected/3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851-kube-api-access-bcmbg\") pod \"memcached-0\" (UID: \"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851\") " pod="openstack/memcached-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.267810 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 04:53:47 crc kubenswrapper[4575]: I1004 04:53:47.342814 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 04:53:48 crc kubenswrapper[4575]: I1004 04:53:48.886003 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:53:48 crc kubenswrapper[4575]: I1004 04:53:48.887252 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:53:48 crc kubenswrapper[4575]: I1004 04:53:48.889908 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-n66ff" Oct 04 04:53:48 crc kubenswrapper[4575]: I1004 04:53:48.961275 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:53:49 crc kubenswrapper[4575]: I1004 04:53:49.054810 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfpvw\" (UniqueName: \"kubernetes.io/projected/010127df-25eb-4251-91a4-5cf00fafc403-kube-api-access-zfpvw\") pod \"kube-state-metrics-0\" (UID: \"010127df-25eb-4251-91a4-5cf00fafc403\") " pod="openstack/kube-state-metrics-0" Oct 04 04:53:49 crc kubenswrapper[4575]: I1004 04:53:49.156351 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfpvw\" (UniqueName: \"kubernetes.io/projected/010127df-25eb-4251-91a4-5cf00fafc403-kube-api-access-zfpvw\") pod \"kube-state-metrics-0\" (UID: \"010127df-25eb-4251-91a4-5cf00fafc403\") " pod="openstack/kube-state-metrics-0" Oct 04 04:53:49 crc kubenswrapper[4575]: I1004 04:53:49.179855 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfpvw\" (UniqueName: \"kubernetes.io/projected/010127df-25eb-4251-91a4-5cf00fafc403-kube-api-access-zfpvw\") pod \"kube-state-metrics-0\" (UID: \"010127df-25eb-4251-91a4-5cf00fafc403\") " pod="openstack/kube-state-metrics-0" Oct 04 04:53:49 crc kubenswrapper[4575]: I1004 04:53:49.257828 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:53:49 crc kubenswrapper[4575]: I1004 04:53:49.735713 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5","Type":"ContainerStarted","Data":"f1297539426409d4a22a7171c35b610bbff4b101ef9949a0c3c3b453a625c0d1"} Oct 04 04:53:52 crc kubenswrapper[4575]: I1004 04:53:52.881744 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 04:53:52 crc kubenswrapper[4575]: I1004 04:53:52.884399 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:52 crc kubenswrapper[4575]: I1004 04:53:52.887374 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 04 04:53:52 crc kubenswrapper[4575]: I1004 04:53:52.887374 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 04 04:53:52 crc kubenswrapper[4575]: I1004 04:53:52.887417 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-wdk6v" Oct 04 04:53:52 crc kubenswrapper[4575]: I1004 04:53:52.887506 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 04 04:53:52 crc kubenswrapper[4575]: I1004 04:53:52.887546 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 04 04:53:52 crc kubenswrapper[4575]: I1004 04:53:52.892227 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.023950 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd50f51c-ef8b-4966-a7d5-3f5420295d17-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.024026 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd50f51c-ef8b-4966-a7d5-3f5420295d17-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.024052 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wf2d\" (UniqueName: \"kubernetes.io/projected/bd50f51c-ef8b-4966-a7d5-3f5420295d17-kube-api-access-5wf2d\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.024141 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd50f51c-ef8b-4966-a7d5-3f5420295d17-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.024261 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd50f51c-ef8b-4966-a7d5-3f5420295d17-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.024336 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.024400 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bd50f51c-ef8b-4966-a7d5-3f5420295d17-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.024424 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd50f51c-ef8b-4966-a7d5-3f5420295d17-config\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.125638 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.125699 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bd50f51c-ef8b-4966-a7d5-3f5420295d17-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.125717 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd50f51c-ef8b-4966-a7d5-3f5420295d17-config\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.125767 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd50f51c-ef8b-4966-a7d5-3f5420295d17-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.125880 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd50f51c-ef8b-4966-a7d5-3f5420295d17-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.126180 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bd50f51c-ef8b-4966-a7d5-3f5420295d17-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.126543 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd50f51c-ef8b-4966-a7d5-3f5420295d17-config\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.125901 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wf2d\" (UniqueName: \"kubernetes.io/projected/bd50f51c-ef8b-4966-a7d5-3f5420295d17-kube-api-access-5wf2d\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.126669 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd50f51c-ef8b-4966-a7d5-3f5420295d17-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.127071 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd50f51c-ef8b-4966-a7d5-3f5420295d17-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.127898 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd50f51c-ef8b-4966-a7d5-3f5420295d17-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.128065 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.140550 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd50f51c-ef8b-4966-a7d5-3f5420295d17-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.141345 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd50f51c-ef8b-4966-a7d5-3f5420295d17-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.150305 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd50f51c-ef8b-4966-a7d5-3f5420295d17-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.183476 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wf2d\" (UniqueName: \"kubernetes.io/projected/bd50f51c-ef8b-4966-a7d5-3f5420295d17-kube-api-access-5wf2d\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.232914 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bd50f51c-ef8b-4966-a7d5-3f5420295d17\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.515332 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.846673 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-95h4j"] Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.852034 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.859184 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.859784 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.901621 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-29972" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.920817 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-95h4j"] Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.942151 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4e95f59-8c21-495d-bc35-151f36e37681-ovn-controller-tls-certs\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.942453 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8wnr\" (UniqueName: \"kubernetes.io/projected/e4e95f59-8c21-495d-bc35-151f36e37681-kube-api-access-l8wnr\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.942644 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4e95f59-8c21-495d-bc35-151f36e37681-var-log-ovn\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.942771 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4e95f59-8c21-495d-bc35-151f36e37681-scripts\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.942875 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4e95f59-8c21-495d-bc35-151f36e37681-combined-ca-bundle\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.943426 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4e95f59-8c21-495d-bc35-151f36e37681-var-run-ovn\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.943547 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4e95f59-8c21-495d-bc35-151f36e37681-var-run\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.946787 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-c66wt"] Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.949308 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:53 crc kubenswrapper[4575]: I1004 04:53:53.959715 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-c66wt"] Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045209 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4e95f59-8c21-495d-bc35-151f36e37681-ovn-controller-tls-certs\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045274 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8wnr\" (UniqueName: \"kubernetes.io/projected/e4e95f59-8c21-495d-bc35-151f36e37681-kube-api-access-l8wnr\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045304 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-etc-ovs\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045331 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4e95f59-8c21-495d-bc35-151f36e37681-var-log-ovn\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045357 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-var-log\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045385 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4e95f59-8c21-495d-bc35-151f36e37681-scripts\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045409 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4e95f59-8c21-495d-bc35-151f36e37681-combined-ca-bundle\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045441 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh78c\" (UniqueName: \"kubernetes.io/projected/1a2a7f61-a13a-4255-83f3-f6c23fc25956-kube-api-access-qh78c\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045469 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4e95f59-8c21-495d-bc35-151f36e37681-var-run-ovn\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045490 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4e95f59-8c21-495d-bc35-151f36e37681-var-run\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045534 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-var-run\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045552 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-var-lib\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.045598 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a2a7f61-a13a-4255-83f3-f6c23fc25956-scripts\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.046464 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4e95f59-8c21-495d-bc35-151f36e37681-var-log-ovn\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.047624 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4e95f59-8c21-495d-bc35-151f36e37681-var-run\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.047852 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4e95f59-8c21-495d-bc35-151f36e37681-var-run-ovn\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.051155 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4e95f59-8c21-495d-bc35-151f36e37681-combined-ca-bundle\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.054662 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4e95f59-8c21-495d-bc35-151f36e37681-scripts\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.067794 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4e95f59-8c21-495d-bc35-151f36e37681-ovn-controller-tls-certs\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.068193 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8wnr\" (UniqueName: \"kubernetes.io/projected/e4e95f59-8c21-495d-bc35-151f36e37681-kube-api-access-l8wnr\") pod \"ovn-controller-95h4j\" (UID: \"e4e95f59-8c21-495d-bc35-151f36e37681\") " pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.147563 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-var-run\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.147943 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-var-lib\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.147987 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a2a7f61-a13a-4255-83f3-f6c23fc25956-scripts\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.148080 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-etc-ovs\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.148104 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-var-log\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.148151 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh78c\" (UniqueName: \"kubernetes.io/projected/1a2a7f61-a13a-4255-83f3-f6c23fc25956-kube-api-access-qh78c\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.147781 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-var-run\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.148766 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-var-lib\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.150117 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-etc-ovs\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.150230 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1a2a7f61-a13a-4255-83f3-f6c23fc25956-var-log\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.151152 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a2a7f61-a13a-4255-83f3-f6c23fc25956-scripts\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.187166 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.188465 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh78c\" (UniqueName: \"kubernetes.io/projected/1a2a7f61-a13a-4255-83f3-f6c23fc25956-kube-api-access-qh78c\") pod \"ovn-controller-ovs-c66wt\" (UID: \"1a2a7f61-a13a-4255-83f3-f6c23fc25956\") " pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:54 crc kubenswrapper[4575]: I1004 04:53:54.276187 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.396144 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.397758 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.401314 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.404752 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-tkfzm" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.405399 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.405760 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.421553 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.492813 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.492868 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ca00752a-2943-4e70-962d-91503f9dd347-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.492894 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffzzh\" (UniqueName: \"kubernetes.io/projected/ca00752a-2943-4e70-962d-91503f9dd347-kube-api-access-ffzzh\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.492939 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca00752a-2943-4e70-962d-91503f9dd347-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.492970 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca00752a-2943-4e70-962d-91503f9dd347-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.493000 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca00752a-2943-4e70-962d-91503f9dd347-config\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.493018 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca00752a-2943-4e70-962d-91503f9dd347-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.493216 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca00752a-2943-4e70-962d-91503f9dd347-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.594959 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca00752a-2943-4e70-962d-91503f9dd347-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.595041 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.595067 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ca00752a-2943-4e70-962d-91503f9dd347-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.595090 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffzzh\" (UniqueName: \"kubernetes.io/projected/ca00752a-2943-4e70-962d-91503f9dd347-kube-api-access-ffzzh\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.595132 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca00752a-2943-4e70-962d-91503f9dd347-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.595163 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca00752a-2943-4e70-962d-91503f9dd347-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.595193 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca00752a-2943-4e70-962d-91503f9dd347-config\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.595213 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca00752a-2943-4e70-962d-91503f9dd347-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.597928 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ca00752a-2943-4e70-962d-91503f9dd347-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.598999 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca00752a-2943-4e70-962d-91503f9dd347-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.599375 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca00752a-2943-4e70-962d-91503f9dd347-config\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.599753 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.611762 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca00752a-2943-4e70-962d-91503f9dd347-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.612425 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca00752a-2943-4e70-962d-91503f9dd347-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.613629 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca00752a-2943-4e70-962d-91503f9dd347-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.615055 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffzzh\" (UniqueName: \"kubernetes.io/projected/ca00752a-2943-4e70-962d-91503f9dd347-kube-api-access-ffzzh\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.623705 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ca00752a-2943-4e70-962d-91503f9dd347\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:56 crc kubenswrapper[4575]: I1004 04:53:56.726695 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 04:53:59 crc kubenswrapper[4575]: I1004 04:53:59.085792 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 04:53:59 crc kubenswrapper[4575]: E1004 04:53:59.821010 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 04:53:59 crc kubenswrapper[4575]: E1004 04:53:59.821181 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xbgh4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-mh9hm_openstack(d2885fac-69b9-4c11-a52c-d88188cee016): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:53:59 crc kubenswrapper[4575]: E1004 04:53:59.822281 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" podUID="d2885fac-69b9-4c11-a52c-d88188cee016" Oct 04 04:53:59 crc kubenswrapper[4575]: E1004 04:53:59.896218 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 04:53:59 crc kubenswrapper[4575]: E1004 04:53:59.896709 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6fnrn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-4wmw2_openstack(6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:53:59 crc kubenswrapper[4575]: E1004 04:53:59.898136 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" podUID="6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.729516 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.739473 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.850283 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-config\") pod \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\" (UID: \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\") " Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.850827 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-config" (OuterVolumeSpecName: "config") pod "6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1" (UID: "6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.851014 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbgh4\" (UniqueName: \"kubernetes.io/projected/d2885fac-69b9-4c11-a52c-d88188cee016-kube-api-access-xbgh4\") pod \"d2885fac-69b9-4c11-a52c-d88188cee016\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.851052 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-config\") pod \"d2885fac-69b9-4c11-a52c-d88188cee016\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.851080 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fnrn\" (UniqueName: \"kubernetes.io/projected/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-kube-api-access-6fnrn\") pod \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\" (UID: \"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1\") " Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.851156 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-dns-svc\") pod \"d2885fac-69b9-4c11-a52c-d88188cee016\" (UID: \"d2885fac-69b9-4c11-a52c-d88188cee016\") " Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.851928 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-config" (OuterVolumeSpecName: "config") pod "d2885fac-69b9-4c11-a52c-d88188cee016" (UID: "d2885fac-69b9-4c11-a52c-d88188cee016"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.856295 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d2885fac-69b9-4c11-a52c-d88188cee016" (UID: "d2885fac-69b9-4c11-a52c-d88188cee016"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.860052 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.860078 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2885fac-69b9-4c11-a52c-d88188cee016-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.860090 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.874393 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-kube-api-access-6fnrn" (OuterVolumeSpecName: "kube-api-access-6fnrn") pod "6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1" (UID: "6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1"). InnerVolumeSpecName "kube-api-access-6fnrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.889761 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2885fac-69b9-4c11-a52c-d88188cee016-kube-api-access-xbgh4" (OuterVolumeSpecName: "kube-api-access-xbgh4") pod "d2885fac-69b9-4c11-a52c-d88188cee016" (UID: "d2885fac-69b9-4c11-a52c-d88188cee016"). InnerVolumeSpecName "kube-api-access-xbgh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.931835 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"04440b88-f2b8-418c-9676-714f0e8b7112","Type":"ContainerStarted","Data":"51842b5fbf5849cb41300944999a624ffd8fe1c05e7ca4e17064f34a87676a32"} Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.935119 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" event={"ID":"6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1","Type":"ContainerDied","Data":"ecf3315f5d2e9510f41d93dd55499ebe0fba989470c7657c59b0405535d419cd"} Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.935210 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4wmw2" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.944983 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" event={"ID":"d2885fac-69b9-4c11-a52c-d88188cee016","Type":"ContainerDied","Data":"1b1ca972adfddf6799e4cc6d9a678ee97ae151b32957f3fd2211b89bbaa296b2"} Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.945093 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mh9hm" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.961682 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbgh4\" (UniqueName: \"kubernetes.io/projected/d2885fac-69b9-4c11-a52c-d88188cee016-kube-api-access-xbgh4\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:01 crc kubenswrapper[4575]: I1004 04:54:01.961713 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fnrn\" (UniqueName: \"kubernetes.io/projected/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1-kube-api-access-6fnrn\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.053436 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4wmw2"] Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.070722 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4wmw2"] Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.083010 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mh9hm"] Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.113631 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mh9hm"] Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.213915 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.248469 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 04:54:02 crc kubenswrapper[4575]: W1004 04:54:02.253843 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c2dcf5d_de87_4624_bd9a_c1b5a2fdc851.slice/crio-ade5d2fe086f17167505893cbf24cc1ab3aea3e08bb8989b36f4b096e17044f2 WatchSource:0}: Error finding container ade5d2fe086f17167505893cbf24cc1ab3aea3e08bb8989b36f4b096e17044f2: Status 404 returned error can't find the container with id ade5d2fe086f17167505893cbf24cc1ab3aea3e08bb8989b36f4b096e17044f2 Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.487059 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.516694 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-95h4j"] Oct 04 04:54:02 crc kubenswrapper[4575]: W1004 04:54:02.597776 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod010127df_25eb_4251_91a4_5cf00fafc403.slice/crio-0d65d6abea2c703a5bd842b96eaf3ca4fb631ec2fc1a22139f16ff853a50ac8e WatchSource:0}: Error finding container 0d65d6abea2c703a5bd842b96eaf3ca4fb631ec2fc1a22139f16ff853a50ac8e: Status 404 returned error can't find the container with id 0d65d6abea2c703a5bd842b96eaf3ca4fb631ec2fc1a22139f16ff853a50ac8e Oct 04 04:54:02 crc kubenswrapper[4575]: W1004 04:54:02.599655 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4e95f59_8c21_495d_bc35_151f36e37681.slice/crio-7a5a03696b8aa6379dbac8f656805704a834fe052087df3c8210b493e07fba49 WatchSource:0}: Error finding container 7a5a03696b8aa6379dbac8f656805704a834fe052087df3c8210b493e07fba49: Status 404 returned error can't find the container with id 7a5a03696b8aa6379dbac8f656805704a834fe052087df3c8210b493e07fba49 Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.961105 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"21199967-fd4f-43af-acd4-653828f2335a","Type":"ContainerStarted","Data":"844f3760e74a1e88edf39c10d3ef420f40c717267d12d038597c58c890e5ecec"} Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.964226 4575 generic.go:334] "Generic (PLEG): container finished" podID="5569b565-1561-4c80-84b4-fa48965c1d1f" containerID="9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b" exitCode=0 Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.964315 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" event={"ID":"5569b565-1561-4c80-84b4-fa48965c1d1f","Type":"ContainerDied","Data":"9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b"} Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.966404 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"010127df-25eb-4251-91a4-5cf00fafc403","Type":"ContainerStarted","Data":"0d65d6abea2c703a5bd842b96eaf3ca4fb631ec2fc1a22139f16ff853a50ac8e"} Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.970574 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-95h4j" event={"ID":"e4e95f59-8c21-495d-bc35-151f36e37681","Type":"ContainerStarted","Data":"7a5a03696b8aa6379dbac8f656805704a834fe052087df3c8210b493e07fba49"} Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.975300 4575 generic.go:334] "Generic (PLEG): container finished" podID="4c489f23-9a31-437f-ae9c-3f421ff5aace" containerID="e8f6fa0a47d77f414f6328b8b0c2a024159998ea065b75a9fdf66df5114b88a4" exitCode=0 Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.976109 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" event={"ID":"4c489f23-9a31-437f-ae9c-3f421ff5aace","Type":"ContainerDied","Data":"e8f6fa0a47d77f414f6328b8b0c2a024159998ea065b75a9fdf66df5114b88a4"} Oct 04 04:54:02 crc kubenswrapper[4575]: I1004 04:54:02.984926 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851","Type":"ContainerStarted","Data":"ade5d2fe086f17167505893cbf24cc1ab3aea3e08bb8989b36f4b096e17044f2"} Oct 04 04:54:03 crc kubenswrapper[4575]: I1004 04:54:03.107319 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 04:54:03 crc kubenswrapper[4575]: W1004 04:54:03.121328 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca00752a_2943_4e70_962d_91503f9dd347.slice/crio-689b0df46d3db271d38476ec7ea8f9191b8738cfffcf57c93fb568ed93431d3d WatchSource:0}: Error finding container 689b0df46d3db271d38476ec7ea8f9191b8738cfffcf57c93fb568ed93431d3d: Status 404 returned error can't find the container with id 689b0df46d3db271d38476ec7ea8f9191b8738cfffcf57c93fb568ed93431d3d Oct 04 04:54:03 crc kubenswrapper[4575]: E1004 04:54:03.263963 4575 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 04 04:54:03 crc kubenswrapper[4575]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/5569b565-1561-4c80-84b4-fa48965c1d1f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 04 04:54:03 crc kubenswrapper[4575]: > podSandboxID="3fc8215aee7adf6d541e3f877b9ae090562b7dc568937edb620e775339a3512b" Oct 04 04:54:03 crc kubenswrapper[4575]: E1004 04:54:03.264157 4575 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 04 04:54:03 crc kubenswrapper[4575]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-46zzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-6cgqb_openstack(5569b565-1561-4c80-84b4-fa48965c1d1f): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/5569b565-1561-4c80-84b4-fa48965c1d1f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 04 04:54:03 crc kubenswrapper[4575]: > logger="UnhandledError" Oct 04 04:54:03 crc kubenswrapper[4575]: E1004 04:54:03.265345 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/5569b565-1561-4c80-84b4-fa48965c1d1f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" podUID="5569b565-1561-4c80-84b4-fa48965c1d1f" Oct 04 04:54:03 crc kubenswrapper[4575]: I1004 04:54:03.340776 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1" path="/var/lib/kubelet/pods/6e7c3d49-73bd-4b2d-9a21-e36b1d307eb1/volumes" Oct 04 04:54:03 crc kubenswrapper[4575]: I1004 04:54:03.341160 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2885fac-69b9-4c11-a52c-d88188cee016" path="/var/lib/kubelet/pods/d2885fac-69b9-4c11-a52c-d88188cee016/volumes" Oct 04 04:54:03 crc kubenswrapper[4575]: I1004 04:54:03.782421 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 04:54:03 crc kubenswrapper[4575]: I1004 04:54:03.942948 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-c66wt"] Oct 04 04:54:03 crc kubenswrapper[4575]: I1004 04:54:03.994946 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c16a4e4c-f361-42ef-9af1-e7601b14ca69","Type":"ContainerStarted","Data":"aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d"} Oct 04 04:54:03 crc kubenswrapper[4575]: I1004 04:54:03.998092 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ca00752a-2943-4e70-962d-91503f9dd347","Type":"ContainerStarted","Data":"689b0df46d3db271d38476ec7ea8f9191b8738cfffcf57c93fb568ed93431d3d"} Oct 04 04:54:04 crc kubenswrapper[4575]: I1004 04:54:04.000139 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5","Type":"ContainerStarted","Data":"b1fa4225952bfbb53d63c7c3c37d6673d544f5c73aa3fa918767cc4e8a9e89ef"} Oct 04 04:54:04 crc kubenswrapper[4575]: I1004 04:54:04.004200 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" event={"ID":"4c489f23-9a31-437f-ae9c-3f421ff5aace","Type":"ContainerStarted","Data":"5dbd852624bba2661f3547c0d8514d697c6e211a17dd4411df1ca73178bd85b9"} Oct 04 04:54:04 crc kubenswrapper[4575]: I1004 04:54:04.004443 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:54:04 crc kubenswrapper[4575]: I1004 04:54:04.076881 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" podStartSLOduration=4.2922672 podStartE2EDuration="22.076836386s" podCreationTimestamp="2025-10-04 04:53:42 +0000 UTC" firstStartedPulling="2025-10-04 04:53:43.922366833 +0000 UTC m=+1175.250925647" lastFinishedPulling="2025-10-04 04:54:01.706936019 +0000 UTC m=+1193.035494833" observedRunningTime="2025-10-04 04:54:04.07350885 +0000 UTC m=+1195.402067664" watchObservedRunningTime="2025-10-04 04:54:04.076836386 +0000 UTC m=+1195.405395230" Oct 04 04:54:06 crc kubenswrapper[4575]: I1004 04:54:06.050481 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bd50f51c-ef8b-4966-a7d5-3f5420295d17","Type":"ContainerStarted","Data":"6410df39a0b9c0894a74b37a0d18cd9a87cf8062eee8eecd719fcf527503c8f2"} Oct 04 04:54:06 crc kubenswrapper[4575]: I1004 04:54:06.051908 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c66wt" event={"ID":"1a2a7f61-a13a-4255-83f3-f6c23fc25956","Type":"ContainerStarted","Data":"318946c4e819aad5508999e14e736bee29b87e09ab7a19c89961ab641f59ee3e"} Oct 04 04:54:08 crc kubenswrapper[4575]: I1004 04:54:08.366828 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:54:08 crc kubenswrapper[4575]: I1004 04:54:08.437046 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6cgqb"] Oct 04 04:54:08 crc kubenswrapper[4575]: I1004 04:54:08.447012 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:54:08 crc kubenswrapper[4575]: I1004 04:54:08.447072 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:54:14 crc kubenswrapper[4575]: E1004 04:54:14.902038 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a2a7f61_a13a_4255_83f3_f6c23fc25956.slice/crio-8a1154029da56799d5cca2009ae02af79f30592f6690ed7b1c5b3a4fa2ebae8a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a2a7f61_a13a_4255_83f3_f6c23fc25956.slice/crio-conmon-8a1154029da56799d5cca2009ae02af79f30592f6690ed7b1c5b3a4fa2ebae8a.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.123421 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bd50f51c-ef8b-4966-a7d5-3f5420295d17","Type":"ContainerStarted","Data":"4e14f40272879b3ce1c2d3b12903dd220f6686206e706e29287f24762e777930"} Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.124978 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"04440b88-f2b8-418c-9676-714f0e8b7112","Type":"ContainerStarted","Data":"1cf9dd2ebebef9c7734fea78e623c981d6fdc16f53daf88cd4510f1c219aae60"} Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.128064 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" event={"ID":"5569b565-1561-4c80-84b4-fa48965c1d1f","Type":"ContainerStarted","Data":"12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290"} Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.128182 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" podUID="5569b565-1561-4c80-84b4-fa48965c1d1f" containerName="dnsmasq-dns" containerID="cri-o://12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290" gracePeriod=10 Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.128306 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.131017 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"010127df-25eb-4251-91a4-5cf00fafc403","Type":"ContainerStarted","Data":"79b8cb3b19aab4a1345eb2ddc54f6a6af93bcc6bdd6b18c48a8aace11cf3b3af"} Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.131117 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.132962 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-95h4j" event={"ID":"e4e95f59-8c21-495d-bc35-151f36e37681","Type":"ContainerStarted","Data":"5a35b4ecb02ac7ef6a18e0ad6a8aad853400c54643d46450ebee68235a362a3a"} Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.133689 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-95h4j" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.135405 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851","Type":"ContainerStarted","Data":"53b8583aabbfed2fc173c73867ffae082362d0d941fb4f7fd9baec795fafe75d"} Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.135933 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.137549 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"21199967-fd4f-43af-acd4-653828f2335a","Type":"ContainerStarted","Data":"17c1336bc9e0f36c1d87a1fbd9dd17a21b8ac0530485311c31d91a1f30f8e82b"} Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.141226 4575 generic.go:334] "Generic (PLEG): container finished" podID="1a2a7f61-a13a-4255-83f3-f6c23fc25956" containerID="8a1154029da56799d5cca2009ae02af79f30592f6690ed7b1c5b3a4fa2ebae8a" exitCode=0 Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.142171 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c66wt" event={"ID":"1a2a7f61-a13a-4255-83f3-f6c23fc25956","Type":"ContainerDied","Data":"8a1154029da56799d5cca2009ae02af79f30592f6690ed7b1c5b3a4fa2ebae8a"} Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.173010 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ca00752a-2943-4e70-962d-91503f9dd347","Type":"ContainerStarted","Data":"80716b82c1ab87292c5bc9529ad54b2b94ed4d27d4b9b886835b1f2f0e4f53d6"} Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.209946 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.879514494 podStartE2EDuration="27.209923023s" podCreationTimestamp="2025-10-04 04:53:48 +0000 UTC" firstStartedPulling="2025-10-04 04:54:02.601796331 +0000 UTC m=+1193.930355145" lastFinishedPulling="2025-10-04 04:54:13.93220487 +0000 UTC m=+1205.260763674" observedRunningTime="2025-10-04 04:54:15.20600989 +0000 UTC m=+1206.534568694" watchObservedRunningTime="2025-10-04 04:54:15.209923023 +0000 UTC m=+1206.538481837" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.233311 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=19.038716677 podStartE2EDuration="29.233288076s" podCreationTimestamp="2025-10-04 04:53:46 +0000 UTC" firstStartedPulling="2025-10-04 04:54:02.258812307 +0000 UTC m=+1193.587371121" lastFinishedPulling="2025-10-04 04:54:12.453383706 +0000 UTC m=+1203.781942520" observedRunningTime="2025-10-04 04:54:15.225179612 +0000 UTC m=+1206.553738436" watchObservedRunningTime="2025-10-04 04:54:15.233288076 +0000 UTC m=+1206.561846900" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.283031 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" podStartSLOduration=15.320353084 podStartE2EDuration="33.283011237s" podCreationTimestamp="2025-10-04 04:53:42 +0000 UTC" firstStartedPulling="2025-10-04 04:53:43.724111786 +0000 UTC m=+1175.052670600" lastFinishedPulling="2025-10-04 04:54:01.686769939 +0000 UTC m=+1193.015328753" observedRunningTime="2025-10-04 04:54:15.276259083 +0000 UTC m=+1206.604817907" watchObservedRunningTime="2025-10-04 04:54:15.283011237 +0000 UTC m=+1206.611570051" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.302529 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-95h4j" podStartSLOduration=11.808762147 podStartE2EDuration="22.302507679s" podCreationTimestamp="2025-10-04 04:53:53 +0000 UTC" firstStartedPulling="2025-10-04 04:54:02.601918825 +0000 UTC m=+1193.930477639" lastFinishedPulling="2025-10-04 04:54:13.095664357 +0000 UTC m=+1204.424223171" observedRunningTime="2025-10-04 04:54:15.297471944 +0000 UTC m=+1206.626030758" watchObservedRunningTime="2025-10-04 04:54:15.302507679 +0000 UTC m=+1206.631066493" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.613367 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.661960 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-config\") pod \"5569b565-1561-4c80-84b4-fa48965c1d1f\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.662108 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-dns-svc\") pod \"5569b565-1561-4c80-84b4-fa48965c1d1f\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.662188 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46zzx\" (UniqueName: \"kubernetes.io/projected/5569b565-1561-4c80-84b4-fa48965c1d1f-kube-api-access-46zzx\") pod \"5569b565-1561-4c80-84b4-fa48965c1d1f\" (UID: \"5569b565-1561-4c80-84b4-fa48965c1d1f\") " Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.670420 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5569b565-1561-4c80-84b4-fa48965c1d1f-kube-api-access-46zzx" (OuterVolumeSpecName: "kube-api-access-46zzx") pod "5569b565-1561-4c80-84b4-fa48965c1d1f" (UID: "5569b565-1561-4c80-84b4-fa48965c1d1f"). InnerVolumeSpecName "kube-api-access-46zzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.746922 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-config" (OuterVolumeSpecName: "config") pod "5569b565-1561-4c80-84b4-fa48965c1d1f" (UID: "5569b565-1561-4c80-84b4-fa48965c1d1f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.756838 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5569b565-1561-4c80-84b4-fa48965c1d1f" (UID: "5569b565-1561-4c80-84b4-fa48965c1d1f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.765242 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.765299 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5569b565-1561-4c80-84b4-fa48965c1d1f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:15 crc kubenswrapper[4575]: I1004 04:54:15.765316 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46zzx\" (UniqueName: \"kubernetes.io/projected/5569b565-1561-4c80-84b4-fa48965c1d1f-kube-api-access-46zzx\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.191872 4575 generic.go:334] "Generic (PLEG): container finished" podID="5569b565-1561-4c80-84b4-fa48965c1d1f" containerID="12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290" exitCode=0 Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.191920 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.191935 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" event={"ID":"5569b565-1561-4c80-84b4-fa48965c1d1f","Type":"ContainerDied","Data":"12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290"} Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.192303 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6cgqb" event={"ID":"5569b565-1561-4c80-84b4-fa48965c1d1f","Type":"ContainerDied","Data":"3fc8215aee7adf6d541e3f877b9ae090562b7dc568937edb620e775339a3512b"} Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.192354 4575 scope.go:117] "RemoveContainer" containerID="12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290" Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.194441 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c66wt" event={"ID":"1a2a7f61-a13a-4255-83f3-f6c23fc25956","Type":"ContainerStarted","Data":"4505651d48711e207fa1a2857d68bf7e31532972db55d90eac809bf5308e951c"} Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.228890 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6cgqb"] Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.235633 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6cgqb"] Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.241722 4575 scope.go:117] "RemoveContainer" containerID="9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b" Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.263790 4575 scope.go:117] "RemoveContainer" containerID="12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290" Oct 04 04:54:16 crc kubenswrapper[4575]: E1004 04:54:16.264643 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290\": container with ID starting with 12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290 not found: ID does not exist" containerID="12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290" Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.264682 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290"} err="failed to get container status \"12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290\": rpc error: code = NotFound desc = could not find container \"12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290\": container with ID starting with 12eedb59addb7a2a3208b6528994fde48557eb685bfe6d0400a8cbdd6fbb0290 not found: ID does not exist" Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.264710 4575 scope.go:117] "RemoveContainer" containerID="9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b" Oct 04 04:54:16 crc kubenswrapper[4575]: E1004 04:54:16.265167 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b\": container with ID starting with 9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b not found: ID does not exist" containerID="9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b" Oct 04 04:54:16 crc kubenswrapper[4575]: I1004 04:54:16.265204 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b"} err="failed to get container status \"9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b\": rpc error: code = NotFound desc = could not find container \"9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b\": container with ID starting with 9bec1d870956932582c2da5893506c84bc6ce2e5596afcaa25766d7683ab679b not found: ID does not exist" Oct 04 04:54:17 crc kubenswrapper[4575]: I1004 04:54:17.203616 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c66wt" event={"ID":"1a2a7f61-a13a-4255-83f3-f6c23fc25956","Type":"ContainerStarted","Data":"f75067a61baac706fbcbab347acef9589d1cca61fab0036febe87ee93f98be65"} Oct 04 04:54:17 crc kubenswrapper[4575]: I1004 04:54:17.204037 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:54:17 crc kubenswrapper[4575]: I1004 04:54:17.204092 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:54:17 crc kubenswrapper[4575]: I1004 04:54:17.226409 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-c66wt" podStartSLOduration=16.794034887 podStartE2EDuration="24.226393675s" podCreationTimestamp="2025-10-04 04:53:53 +0000 UTC" firstStartedPulling="2025-10-04 04:54:05.806788369 +0000 UTC m=+1197.135347183" lastFinishedPulling="2025-10-04 04:54:13.239147157 +0000 UTC m=+1204.567705971" observedRunningTime="2025-10-04 04:54:17.226142128 +0000 UTC m=+1208.554700952" watchObservedRunningTime="2025-10-04 04:54:17.226393675 +0000 UTC m=+1208.554952489" Oct 04 04:54:17 crc kubenswrapper[4575]: I1004 04:54:17.321279 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5569b565-1561-4c80-84b4-fa48965c1d1f" path="/var/lib/kubelet/pods/5569b565-1561-4c80-84b4-fa48965c1d1f/volumes" Oct 04 04:54:19 crc kubenswrapper[4575]: I1004 04:54:19.262201 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 04:54:22 crc kubenswrapper[4575]: I1004 04:54:22.344696 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 04 04:54:29 crc kubenswrapper[4575]: I1004 04:54:29.973208 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hd5w5"] Oct 04 04:54:29 crc kubenswrapper[4575]: E1004 04:54:29.978681 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5569b565-1561-4c80-84b4-fa48965c1d1f" containerName="init" Oct 04 04:54:29 crc kubenswrapper[4575]: I1004 04:54:29.978768 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="5569b565-1561-4c80-84b4-fa48965c1d1f" containerName="init" Oct 04 04:54:29 crc kubenswrapper[4575]: E1004 04:54:29.978852 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5569b565-1561-4c80-84b4-fa48965c1d1f" containerName="dnsmasq-dns" Oct 04 04:54:29 crc kubenswrapper[4575]: I1004 04:54:29.978926 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="5569b565-1561-4c80-84b4-fa48965c1d1f" containerName="dnsmasq-dns" Oct 04 04:54:29 crc kubenswrapper[4575]: I1004 04:54:29.979180 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="5569b565-1561-4c80-84b4-fa48965c1d1f" containerName="dnsmasq-dns" Oct 04 04:54:29 crc kubenswrapper[4575]: I1004 04:54:29.980933 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.025868 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hd5w5"] Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.095811 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-hd5w5\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.095849 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-config\") pod \"dnsmasq-dns-7cb5889db5-hd5w5\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.095907 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frkkl\" (UniqueName: \"kubernetes.io/projected/a94cdf5d-d25e-4806-aec2-8a3958118c6f-kube-api-access-frkkl\") pod \"dnsmasq-dns-7cb5889db5-hd5w5\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.197689 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-config\") pod \"dnsmasq-dns-7cb5889db5-hd5w5\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.197801 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frkkl\" (UniqueName: \"kubernetes.io/projected/a94cdf5d-d25e-4806-aec2-8a3958118c6f-kube-api-access-frkkl\") pod \"dnsmasq-dns-7cb5889db5-hd5w5\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.197938 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-hd5w5\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.199089 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-hd5w5\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.199118 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-config\") pod \"dnsmasq-dns-7cb5889db5-hd5w5\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.223829 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frkkl\" (UniqueName: \"kubernetes.io/projected/a94cdf5d-d25e-4806-aec2-8a3958118c6f-kube-api-access-frkkl\") pod \"dnsmasq-dns-7cb5889db5-hd5w5\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:30 crc kubenswrapper[4575]: I1004 04:54:30.315147 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.339332 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.345830 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.352391 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.352663 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.352987 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-d9wqc" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.353150 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.368680 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.417331 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wr2n\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-kube-api-access-9wr2n\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.417391 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c9ec5625-8b15-42bd-8258-4960d8000469-lock\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.417499 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.417534 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c9ec5625-8b15-42bd-8258-4960d8000469-cache\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.417687 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.518644 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.518730 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wr2n\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-kube-api-access-9wr2n\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.518759 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c9ec5625-8b15-42bd-8258-4960d8000469-lock\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.518834 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.518859 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c9ec5625-8b15-42bd-8258-4960d8000469-cache\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: E1004 04:54:31.518887 4575 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:54:31 crc kubenswrapper[4575]: E1004 04:54:31.518919 4575 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:54:31 crc kubenswrapper[4575]: E1004 04:54:31.519002 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift podName:c9ec5625-8b15-42bd-8258-4960d8000469 nodeName:}" failed. No retries permitted until 2025-10-04 04:54:32.01895395 +0000 UTC m=+1223.347512764 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift") pod "swift-storage-0" (UID: "c9ec5625-8b15-42bd-8258-4960d8000469") : configmap "swift-ring-files" not found Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.519310 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c9ec5625-8b15-42bd-8258-4960d8000469-cache\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.519447 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c9ec5625-8b15-42bd-8258-4960d8000469-lock\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.519528 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.538506 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wr2n\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-kube-api-access-9wr2n\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:31 crc kubenswrapper[4575]: I1004 04:54:31.541173 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:32 crc kubenswrapper[4575]: I1004 04:54:32.027033 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:32 crc kubenswrapper[4575]: E1004 04:54:32.027203 4575 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:54:32 crc kubenswrapper[4575]: E1004 04:54:32.027225 4575 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:54:32 crc kubenswrapper[4575]: E1004 04:54:32.027274 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift podName:c9ec5625-8b15-42bd-8258-4960d8000469 nodeName:}" failed. No retries permitted until 2025-10-04 04:54:33.027259464 +0000 UTC m=+1224.355818278 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift") pod "swift-storage-0" (UID: "c9ec5625-8b15-42bd-8258-4960d8000469") : configmap "swift-ring-files" not found Oct 04 04:54:33 crc kubenswrapper[4575]: I1004 04:54:33.041795 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:33 crc kubenswrapper[4575]: E1004 04:54:33.042014 4575 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:54:33 crc kubenswrapper[4575]: E1004 04:54:33.042035 4575 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:54:33 crc kubenswrapper[4575]: E1004 04:54:33.042136 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift podName:c9ec5625-8b15-42bd-8258-4960d8000469 nodeName:}" failed. No retries permitted until 2025-10-04 04:54:35.04211547 +0000 UTC m=+1226.370674284 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift") pod "swift-storage-0" (UID: "c9ec5625-8b15-42bd-8258-4960d8000469") : configmap "swift-ring-files" not found Oct 04 04:54:34 crc kubenswrapper[4575]: I1004 04:54:34.972995 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-jw946"] Oct 04 04:54:34 crc kubenswrapper[4575]: I1004 04:54:34.974303 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:34 crc kubenswrapper[4575]: I1004 04:54:34.976954 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 04 04:54:34 crc kubenswrapper[4575]: I1004 04:54:34.977287 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 04 04:54:34 crc kubenswrapper[4575]: I1004 04:54:34.977449 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 04:54:34 crc kubenswrapper[4575]: I1004 04:54:34.995424 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jw946"] Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.069825 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-scripts\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.070151 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-dispersionconf\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.070263 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5724856-06a7-4b89-a67e-8239308e798f-etc-swift\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.070395 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-ring-data-devices\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.070511 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmzdb\" (UniqueName: \"kubernetes.io/projected/d5724856-06a7-4b89-a67e-8239308e798f-kube-api-access-tmzdb\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.070597 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-combined-ca-bundle\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.070705 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.070804 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-swiftconf\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: E1004 04:54:35.070891 4575 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:54:35 crc kubenswrapper[4575]: E1004 04:54:35.070919 4575 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:54:35 crc kubenswrapper[4575]: E1004 04:54:35.070964 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift podName:c9ec5625-8b15-42bd-8258-4960d8000469 nodeName:}" failed. No retries permitted until 2025-10-04 04:54:39.070949307 +0000 UTC m=+1230.399508181 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift") pod "swift-storage-0" (UID: "c9ec5625-8b15-42bd-8258-4960d8000469") : configmap "swift-ring-files" not found Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.171951 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmzdb\" (UniqueName: \"kubernetes.io/projected/d5724856-06a7-4b89-a67e-8239308e798f-kube-api-access-tmzdb\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.172003 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-combined-ca-bundle\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.172060 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-swiftconf\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.172096 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-scripts\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.172149 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-dispersionconf\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.172187 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5724856-06a7-4b89-a67e-8239308e798f-etc-swift\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.172228 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-ring-data-devices\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.173391 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-ring-data-devices\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.173472 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-scripts\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.174729 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5724856-06a7-4b89-a67e-8239308e798f-etc-swift\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.185717 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-swiftconf\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.185820 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-combined-ca-bundle\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.185988 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-dispersionconf\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.193543 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmzdb\" (UniqueName: \"kubernetes.io/projected/d5724856-06a7-4b89-a67e-8239308e798f-kube-api-access-tmzdb\") pod \"swift-ring-rebalance-jw946\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.289308 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.337217 4575 generic.go:334] "Generic (PLEG): container finished" podID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" containerID="aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d" exitCode=0 Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.337293 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c16a4e4c-f361-42ef-9af1-e7601b14ca69","Type":"ContainerDied","Data":"aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d"} Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.346035 4575 generic.go:334] "Generic (PLEG): container finished" podID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" containerID="b1fa4225952bfbb53d63c7c3c37d6673d544f5c73aa3fa918767cc4e8a9e89ef" exitCode=0 Oct 04 04:54:35 crc kubenswrapper[4575]: I1004 04:54:35.346077 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5","Type":"ContainerDied","Data":"b1fa4225952bfbb53d63c7c3c37d6673d544f5c73aa3fa918767cc4e8a9e89ef"} Oct 04 04:54:35 crc kubenswrapper[4575]: E1004 04:54:35.348571 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc16a4e4c_f361_42ef_9af1_e7601b14ca69.slice/crio-aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:54:36 crc kubenswrapper[4575]: E1004 04:54:36.522625 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Oct 04 04:54:36 crc kubenswrapper[4575]: E1004 04:54:36.523187 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffzzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(ca00752a-2943-4e70-962d-91503f9dd347): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:54:36 crc kubenswrapper[4575]: E1004 04:54:36.524557 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="ca00752a-2943-4e70-962d-91503f9dd347" Oct 04 04:54:36 crc kubenswrapper[4575]: E1004 04:54:36.616577 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Oct 04 04:54:36 crc kubenswrapper[4575]: E1004 04:54:36.617344 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5wf2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(bd50f51c-ef8b-4966-a7d5-3f5420295d17): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:54:36 crc kubenswrapper[4575]: E1004 04:54:36.618749 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="bd50f51c-ef8b-4966-a7d5-3f5420295d17" Oct 04 04:54:36 crc kubenswrapper[4575]: I1004 04:54:36.976935 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jw946"] Oct 04 04:54:36 crc kubenswrapper[4575]: W1004 04:54:36.979812 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5724856_06a7_4b89_a67e_8239308e798f.slice/crio-f809bffca6509f39aa7b74c4e2c41817cb4d11a7d0367dcae01a63b554c4fdcd WatchSource:0}: Error finding container f809bffca6509f39aa7b74c4e2c41817cb4d11a7d0367dcae01a63b554c4fdcd: Status 404 returned error can't find the container with id f809bffca6509f39aa7b74c4e2c41817cb4d11a7d0367dcae01a63b554c4fdcd Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.058680 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hd5w5"] Oct 04 04:54:37 crc kubenswrapper[4575]: W1004 04:54:37.069647 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda94cdf5d_d25e_4806_aec2_8a3958118c6f.slice/crio-45720c84bacf4da18a311f28185ca4702f5abb113e4d703a382d6bce7678f682 WatchSource:0}: Error finding container 45720c84bacf4da18a311f28185ca4702f5abb113e4d703a382d6bce7678f682: Status 404 returned error can't find the container with id 45720c84bacf4da18a311f28185ca4702f5abb113e4d703a382d6bce7678f682 Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.359450 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jw946" event={"ID":"d5724856-06a7-4b89-a67e-8239308e798f","Type":"ContainerStarted","Data":"f809bffca6509f39aa7b74c4e2c41817cb4d11a7d0367dcae01a63b554c4fdcd"} Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.361410 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5","Type":"ContainerStarted","Data":"2d02ce72a5f63bcbca2e3120fa4a39e38e4cfa8d01de232b93c3fde2b1e34412"} Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.362388 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.363806 4575 generic.go:334] "Generic (PLEG): container finished" podID="21199967-fd4f-43af-acd4-653828f2335a" containerID="17c1336bc9e0f36c1d87a1fbd9dd17a21b8ac0530485311c31d91a1f30f8e82b" exitCode=0 Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.363865 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"21199967-fd4f-43af-acd4-653828f2335a","Type":"ContainerDied","Data":"17c1336bc9e0f36c1d87a1fbd9dd17a21b8ac0530485311c31d91a1f30f8e82b"} Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.367466 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c16a4e4c-f361-42ef-9af1-e7601b14ca69","Type":"ContainerStarted","Data":"827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717"} Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.367857 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.372367 4575 generic.go:334] "Generic (PLEG): container finished" podID="a94cdf5d-d25e-4806-aec2-8a3958118c6f" containerID="1538fe6ea425f8567cf277bed33148cb0995e7a917312034588efbb966823210" exitCode=0 Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.372426 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" event={"ID":"a94cdf5d-d25e-4806-aec2-8a3958118c6f","Type":"ContainerDied","Data":"1538fe6ea425f8567cf277bed33148cb0995e7a917312034588efbb966823210"} Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.372448 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" event={"ID":"a94cdf5d-d25e-4806-aec2-8a3958118c6f","Type":"ContainerStarted","Data":"45720c84bacf4da18a311f28185ca4702f5abb113e4d703a382d6bce7678f682"} Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.375824 4575 generic.go:334] "Generic (PLEG): container finished" podID="04440b88-f2b8-418c-9676-714f0e8b7112" containerID="1cf9dd2ebebef9c7734fea78e623c981d6fdc16f53daf88cd4510f1c219aae60" exitCode=0 Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.376064 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"04440b88-f2b8-418c-9676-714f0e8b7112","Type":"ContainerDied","Data":"1cf9dd2ebebef9c7734fea78e623c981d6fdc16f53daf88cd4510f1c219aae60"} Oct 04 04:54:37 crc kubenswrapper[4575]: E1004 04:54:37.378249 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="ca00752a-2943-4e70-962d-91503f9dd347" Oct 04 04:54:37 crc kubenswrapper[4575]: E1004 04:54:37.382166 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bd50f51c-ef8b-4966-a7d5-3f5420295d17" Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.394452 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.509466326 podStartE2EDuration="54.394431427s" podCreationTimestamp="2025-10-04 04:53:43 +0000 UTC" firstStartedPulling="2025-10-04 04:53:48.919962959 +0000 UTC m=+1180.248521773" lastFinishedPulling="2025-10-04 04:54:01.80492806 +0000 UTC m=+1193.133486874" observedRunningTime="2025-10-04 04:54:37.390760341 +0000 UTC m=+1228.719319165" watchObservedRunningTime="2025-10-04 04:54:37.394431427 +0000 UTC m=+1228.722990251" Oct 04 04:54:37 crc kubenswrapper[4575]: I1004 04:54:37.505957 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.4896598 podStartE2EDuration="55.505932017s" podCreationTimestamp="2025-10-04 04:53:42 +0000 UTC" firstStartedPulling="2025-10-04 04:53:44.684623408 +0000 UTC m=+1176.013182222" lastFinishedPulling="2025-10-04 04:54:01.700895625 +0000 UTC m=+1193.029454439" observedRunningTime="2025-10-04 04:54:37.449023458 +0000 UTC m=+1228.777582282" watchObservedRunningTime="2025-10-04 04:54:37.505932017 +0000 UTC m=+1228.834490841" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.395248 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"21199967-fd4f-43af-acd4-653828f2335a","Type":"ContainerStarted","Data":"602e5bfe7ff10e27575cd085cd8a3fa55b8152796e1a79317947e0250c779ae8"} Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.398902 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" event={"ID":"a94cdf5d-d25e-4806-aec2-8a3958118c6f","Type":"ContainerStarted","Data":"350ab4544bb2bf7d410dfb2899c2d7e618fde475724741bbbe36a2ff1009972f"} Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.399076 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.400774 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"04440b88-f2b8-418c-9676-714f0e8b7112","Type":"ContainerStarted","Data":"1ce8bef38bef2e4637393989b4de9158f060356e4200ef1e3647e5d7f41e6c68"} Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.448630 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.448885 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.449035 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.450119 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ed1625c85b01a1f2dd9735e9997d0b74d3638c1dc9b14bc0e29139e2f4da8d2"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.450252 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://3ed1625c85b01a1f2dd9735e9997d0b74d3638c1dc9b14bc0e29139e2f4da8d2" gracePeriod=600 Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.485256 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=42.495403217 podStartE2EDuration="53.48523483s" podCreationTimestamp="2025-10-04 04:53:45 +0000 UTC" firstStartedPulling="2025-10-04 04:54:02.250806197 +0000 UTC m=+1193.579365011" lastFinishedPulling="2025-10-04 04:54:13.24063782 +0000 UTC m=+1204.569196624" observedRunningTime="2025-10-04 04:54:38.441863591 +0000 UTC m=+1229.770422405" watchObservedRunningTime="2025-10-04 04:54:38.48523483 +0000 UTC m=+1229.813793664" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.516625 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.517241 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 04 04:54:38 crc kubenswrapper[4575]: E1004 04:54:38.519638 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bd50f51c-ef8b-4966-a7d5-3f5420295d17" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.550090 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" podStartSLOduration=9.550069066 podStartE2EDuration="9.550069066s" podCreationTimestamp="2025-10-04 04:54:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:38.546564115 +0000 UTC m=+1229.875122939" watchObservedRunningTime="2025-10-04 04:54:38.550069066 +0000 UTC m=+1229.878627890" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.559329 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=41.785611411 podStartE2EDuration="53.559309702s" podCreationTimestamp="2025-10-04 04:53:45 +0000 UTC" firstStartedPulling="2025-10-04 04:54:01.545838131 +0000 UTC m=+1192.874396945" lastFinishedPulling="2025-10-04 04:54:13.319536422 +0000 UTC m=+1204.648095236" observedRunningTime="2025-10-04 04:54:38.496902806 +0000 UTC m=+1229.825461640" watchObservedRunningTime="2025-10-04 04:54:38.559309702 +0000 UTC m=+1229.887868526" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.575218 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.728106 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 04 04:54:38 crc kubenswrapper[4575]: E1004 04:54:38.730898 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="ca00752a-2943-4e70-962d-91503f9dd347" Oct 04 04:54:38 crc kubenswrapper[4575]: I1004 04:54:38.840548 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 04 04:54:39 crc kubenswrapper[4575]: I1004 04:54:39.160795 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:39 crc kubenswrapper[4575]: E1004 04:54:39.161032 4575 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:54:39 crc kubenswrapper[4575]: E1004 04:54:39.161054 4575 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:54:39 crc kubenswrapper[4575]: E1004 04:54:39.161115 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift podName:c9ec5625-8b15-42bd-8258-4960d8000469 nodeName:}" failed. No retries permitted until 2025-10-04 04:54:47.161091837 +0000 UTC m=+1238.489650651 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift") pod "swift-storage-0" (UID: "c9ec5625-8b15-42bd-8258-4960d8000469") : configmap "swift-ring-files" not found Oct 04 04:54:39 crc kubenswrapper[4575]: I1004 04:54:39.414059 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="3ed1625c85b01a1f2dd9735e9997d0b74d3638c1dc9b14bc0e29139e2f4da8d2" exitCode=0 Oct 04 04:54:39 crc kubenswrapper[4575]: I1004 04:54:39.415075 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"3ed1625c85b01a1f2dd9735e9997d0b74d3638c1dc9b14bc0e29139e2f4da8d2"} Oct 04 04:54:39 crc kubenswrapper[4575]: I1004 04:54:39.415120 4575 scope.go:117] "RemoveContainer" containerID="e6585e9613991e3a385e6fad309ba0255360c49e2f3c2664f8dc4f993639cb30" Oct 04 04:54:39 crc kubenswrapper[4575]: I1004 04:54:39.416314 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 04 04:54:39 crc kubenswrapper[4575]: E1004 04:54:39.416477 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="ca00752a-2943-4e70-962d-91503f9dd347" Oct 04 04:54:39 crc kubenswrapper[4575]: E1004 04:54:39.416916 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bd50f51c-ef8b-4966-a7d5-3f5420295d17" Oct 04 04:54:39 crc kubenswrapper[4575]: I1004 04:54:39.480316 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 04 04:54:39 crc kubenswrapper[4575]: I1004 04:54:39.496709 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 04 04:54:40 crc kubenswrapper[4575]: E1004 04:54:40.428767 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="ca00752a-2943-4e70-962d-91503f9dd347" Oct 04 04:54:40 crc kubenswrapper[4575]: E1004 04:54:40.428882 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bd50f51c-ef8b-4966-a7d5-3f5420295d17" Oct 04 04:54:41 crc kubenswrapper[4575]: E1004 04:54:41.436499 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="ca00752a-2943-4e70-962d-91503f9dd347" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.687263 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-xpsnf"] Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.688326 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.693395 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.705721 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xpsnf"] Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.804698 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-ovs-rundir\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.804788 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.804836 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-884wx\" (UniqueName: \"kubernetes.io/projected/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-kube-api-access-884wx\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.804909 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-combined-ca-bundle\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.805008 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-ovn-rundir\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.805035 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-config\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.907486 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-combined-ca-bundle\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.907616 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-ovn-rundir\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.907652 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-config\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.907701 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-ovs-rundir\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.907775 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.907806 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-884wx\" (UniqueName: \"kubernetes.io/projected/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-kube-api-access-884wx\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.914162 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-combined-ca-bundle\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.915226 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-ovn-rundir\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.916047 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-config\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.916110 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-ovs-rundir\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.938953 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:41 crc kubenswrapper[4575]: I1004 04:54:41.944986 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-884wx\" (UniqueName: \"kubernetes.io/projected/ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3-kube-api-access-884wx\") pod \"ovn-controller-metrics-xpsnf\" (UID: \"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3\") " pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.015996 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xpsnf" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.198806 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hd5w5"] Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.199077 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" podUID="a94cdf5d-d25e-4806-aec2-8a3958118c6f" containerName="dnsmasq-dns" containerID="cri-o://350ab4544bb2bf7d410dfb2899c2d7e618fde475724741bbbe36a2ff1009972f" gracePeriod=10 Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.202728 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.313012 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-qntzr"] Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.314732 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: W1004 04:54:42.322871 4575 reflector.go:561] object-"openstack"/"ovsdbserver-nb": failed to list *v1.ConfigMap: configmaps "ovsdbserver-nb" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 04 04:54:42 crc kubenswrapper[4575]: E1004 04:54:42.322919 4575 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ovsdbserver-nb\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovsdbserver-nb\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.425394 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-dns-svc\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.425812 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-config\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.425847 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxtbn\" (UniqueName: \"kubernetes.io/projected/08b75e8f-7a43-44e6-9d75-24d45fae776b-kube-api-access-qxtbn\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.425899 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.426272 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-qntzr"] Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.465955 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jw946" event={"ID":"d5724856-06a7-4b89-a67e-8239308e798f","Type":"ContainerStarted","Data":"e3c14dbe8ba059270e8e4625059100fd4118169cac6e704e71a896c5c3ab8256"} Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.488042 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"fec7eca9bd28e99b532e638e857f42e7eb7fe8c1e89c0dd52e522352fd86d19d"} Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.530357 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-config\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.530413 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxtbn\" (UniqueName: \"kubernetes.io/projected/08b75e8f-7a43-44e6-9d75-24d45fae776b-kube-api-access-qxtbn\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.539080 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-config\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.540252 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.540487 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-dns-svc\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.552044 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-dns-svc\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.574429 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-jw946" podStartSLOduration=4.19555235 podStartE2EDuration="8.574409032s" podCreationTimestamp="2025-10-04 04:54:34 +0000 UTC" firstStartedPulling="2025-10-04 04:54:36.981995003 +0000 UTC m=+1228.310553817" lastFinishedPulling="2025-10-04 04:54:41.360851685 +0000 UTC m=+1232.689410499" observedRunningTime="2025-10-04 04:54:42.571960592 +0000 UTC m=+1233.900519406" watchObservedRunningTime="2025-10-04 04:54:42.574409032 +0000 UTC m=+1233.902967856" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.606782 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxtbn\" (UniqueName: \"kubernetes.io/projected/08b75e8f-7a43-44e6-9d75-24d45fae776b-kube-api-access-qxtbn\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.809255 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-qntzr"] Oct 04 04:54:42 crc kubenswrapper[4575]: E1004 04:54:42.810252 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-57d65f699f-qntzr" podUID="08b75e8f-7a43-44e6-9d75-24d45fae776b" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.858094 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gm5df"] Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.859856 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.866943 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 04 04:54:42 crc kubenswrapper[4575]: I1004 04:54:42.896311 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gm5df"] Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.043352 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xpsnf"] Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.051010 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.051057 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.051127 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-config\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.051242 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xll4s\" (UniqueName: \"kubernetes.io/projected/12365747-76ba-4bcb-a828-147712279d49-kube-api-access-xll4s\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.051276 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.152500 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xll4s\" (UniqueName: \"kubernetes.io/projected/12365747-76ba-4bcb-a828-147712279d49-kube-api-access-xll4s\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.152877 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.153032 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.153168 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.153328 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-config\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.153916 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.154066 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.154739 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-config\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.177420 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xll4s\" (UniqueName: \"kubernetes.io/projected/12365747-76ba-4bcb-a828-147712279d49-kube-api-access-xll4s\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.499668 4575 generic.go:334] "Generic (PLEG): container finished" podID="a94cdf5d-d25e-4806-aec2-8a3958118c6f" containerID="350ab4544bb2bf7d410dfb2899c2d7e618fde475724741bbbe36a2ff1009972f" exitCode=0 Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.499752 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" event={"ID":"a94cdf5d-d25e-4806-aec2-8a3958118c6f","Type":"ContainerDied","Data":"350ab4544bb2bf7d410dfb2899c2d7e618fde475724741bbbe36a2ff1009972f"} Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.506526 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.506545 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xpsnf" event={"ID":"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3","Type":"ContainerStarted","Data":"03ded492f04de8713654aa1ef15796018f4573472af555a9fe4551b4ee8996ec"} Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.521384 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:43 crc kubenswrapper[4575]: E1004 04:54:43.540644 4575 configmap.go:193] Couldn't get configMap openstack/ovsdbserver-nb: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:54:43 crc kubenswrapper[4575]: E1004 04:54:43.540727 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-ovsdbserver-nb podName:08b75e8f-7a43-44e6-9d75-24d45fae776b nodeName:}" failed. No retries permitted until 2025-10-04 04:54:44.040709911 +0000 UTC m=+1235.369268715 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovsdbserver-nb" (UniqueName: "kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-ovsdbserver-nb") pod "dnsmasq-dns-57d65f699f-qntzr" (UID: "08b75e8f-7a43-44e6-9d75-24d45fae776b") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.660769 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-dns-svc\") pod \"08b75e8f-7a43-44e6-9d75-24d45fae776b\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.661172 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxtbn\" (UniqueName: \"kubernetes.io/projected/08b75e8f-7a43-44e6-9d75-24d45fae776b-kube-api-access-qxtbn\") pod \"08b75e8f-7a43-44e6-9d75-24d45fae776b\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.661218 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08b75e8f-7a43-44e6-9d75-24d45fae776b" (UID: "08b75e8f-7a43-44e6-9d75-24d45fae776b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.661238 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-config\") pod \"08b75e8f-7a43-44e6-9d75-24d45fae776b\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.661608 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-config" (OuterVolumeSpecName: "config") pod "08b75e8f-7a43-44e6-9d75-24d45fae776b" (UID: "08b75e8f-7a43-44e6-9d75-24d45fae776b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.662075 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.662095 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.689978 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b75e8f-7a43-44e6-9d75-24d45fae776b-kube-api-access-qxtbn" (OuterVolumeSpecName: "kube-api-access-qxtbn") pod "08b75e8f-7a43-44e6-9d75-24d45fae776b" (UID: "08b75e8f-7a43-44e6-9d75-24d45fae776b"). InnerVolumeSpecName "kube-api-access-qxtbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.768213 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxtbn\" (UniqueName: \"kubernetes.io/projected/08b75e8f-7a43-44e6-9d75-24d45fae776b-kube-api-access-qxtbn\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.865297 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.874474 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-gm5df\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:43 crc kubenswrapper[4575]: I1004 04:54:43.919092 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.071761 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-config\") pod \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.071882 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frkkl\" (UniqueName: \"kubernetes.io/projected/a94cdf5d-d25e-4806-aec2-8a3958118c6f-kube-api-access-frkkl\") pod \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.071986 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-dns-svc\") pod \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\" (UID: \"a94cdf5d-d25e-4806-aec2-8a3958118c6f\") " Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.072235 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.073088 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-qntzr\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.077675 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a94cdf5d-d25e-4806-aec2-8a3958118c6f-kube-api-access-frkkl" (OuterVolumeSpecName: "kube-api-access-frkkl") pod "a94cdf5d-d25e-4806-aec2-8a3958118c6f" (UID: "a94cdf5d-d25e-4806-aec2-8a3958118c6f"). InnerVolumeSpecName "kube-api-access-frkkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.088655 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.135103 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-config" (OuterVolumeSpecName: "config") pod "a94cdf5d-d25e-4806-aec2-8a3958118c6f" (UID: "a94cdf5d-d25e-4806-aec2-8a3958118c6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.144767 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a94cdf5d-d25e-4806-aec2-8a3958118c6f" (UID: "a94cdf5d-d25e-4806-aec2-8a3958118c6f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.173428 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-ovsdbserver-nb\") pod \"08b75e8f-7a43-44e6-9d75-24d45fae776b\" (UID: \"08b75e8f-7a43-44e6-9d75-24d45fae776b\") " Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.173789 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "08b75e8f-7a43-44e6-9d75-24d45fae776b" (UID: "08b75e8f-7a43-44e6-9d75-24d45fae776b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.174409 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.174433 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08b75e8f-7a43-44e6-9d75-24d45fae776b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.174444 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a94cdf5d-d25e-4806-aec2-8a3958118c6f-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.174455 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frkkl\" (UniqueName: \"kubernetes.io/projected/a94cdf5d-d25e-4806-aec2-8a3958118c6f-kube-api-access-frkkl\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.241869 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-95h4j" podUID="e4e95f59-8c21-495d-bc35-151f36e37681" containerName="ovn-controller" probeResult="failure" output=< Oct 04 04:54:44 crc kubenswrapper[4575]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 04:54:44 crc kubenswrapper[4575]: > Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.479624 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gm5df"] Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.526829 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" event={"ID":"a94cdf5d-d25e-4806-aec2-8a3958118c6f","Type":"ContainerDied","Data":"45720c84bacf4da18a311f28185ca4702f5abb113e4d703a382d6bce7678f682"} Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.526879 4575 scope.go:117] "RemoveContainer" containerID="350ab4544bb2bf7d410dfb2899c2d7e618fde475724741bbbe36a2ff1009972f" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.526997 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-hd5w5" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.545483 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-qntzr" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.547535 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" event={"ID":"12365747-76ba-4bcb-a828-147712279d49","Type":"ContainerStarted","Data":"56e379fc5f2b32ddca9caf1e1330ee1fe92f1360b64525e2318ad2306b9c6d6a"} Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.576198 4575 scope.go:117] "RemoveContainer" containerID="1538fe6ea425f8567cf277bed33148cb0995e7a917312034588efbb966823210" Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.611454 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hd5w5"] Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.617289 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hd5w5"] Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.648124 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-qntzr"] Oct 04 04:54:44 crc kubenswrapper[4575]: I1004 04:54:44.669596 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-qntzr"] Oct 04 04:54:45 crc kubenswrapper[4575]: I1004 04:54:45.321616 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b75e8f-7a43-44e6-9d75-24d45fae776b" path="/var/lib/kubelet/pods/08b75e8f-7a43-44e6-9d75-24d45fae776b/volumes" Oct 04 04:54:45 crc kubenswrapper[4575]: I1004 04:54:45.322371 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a94cdf5d-d25e-4806-aec2-8a3958118c6f" path="/var/lib/kubelet/pods/a94cdf5d-d25e-4806-aec2-8a3958118c6f/volumes" Oct 04 04:54:45 crc kubenswrapper[4575]: I1004 04:54:45.553876 4575 generic.go:334] "Generic (PLEG): container finished" podID="12365747-76ba-4bcb-a828-147712279d49" containerID="fcf6b1710d40d50b14f172a04117c66e789adca0a5d14568ebc089c43098dc71" exitCode=0 Oct 04 04:54:45 crc kubenswrapper[4575]: I1004 04:54:45.553914 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" event={"ID":"12365747-76ba-4bcb-a828-147712279d49","Type":"ContainerDied","Data":"fcf6b1710d40d50b14f172a04117c66e789adca0a5d14568ebc089c43098dc71"} Oct 04 04:54:45 crc kubenswrapper[4575]: E1004 04:54:45.587599 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12365747_76ba_4bcb_a828_147712279d49.slice/crio-conmon-fcf6b1710d40d50b14f172a04117c66e789adca0a5d14568ebc089c43098dc71.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:54:46 crc kubenswrapper[4575]: I1004 04:54:46.554364 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 04 04:54:46 crc kubenswrapper[4575]: I1004 04:54:46.554678 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 04 04:54:47 crc kubenswrapper[4575]: I1004 04:54:47.240962 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:54:47 crc kubenswrapper[4575]: E1004 04:54:47.241296 4575 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:54:47 crc kubenswrapper[4575]: E1004 04:54:47.241451 4575 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:54:47 crc kubenswrapper[4575]: E1004 04:54:47.241503 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift podName:c9ec5625-8b15-42bd-8258-4960d8000469 nodeName:}" failed. No retries permitted until 2025-10-04 04:55:03.241487482 +0000 UTC m=+1254.570046296 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift") pod "swift-storage-0" (UID: "c9ec5625-8b15-42bd-8258-4960d8000469") : configmap "swift-ring-files" not found Oct 04 04:54:47 crc kubenswrapper[4575]: I1004 04:54:47.273286 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 04 04:54:47 crc kubenswrapper[4575]: I1004 04:54:47.274566 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 04 04:54:48 crc kubenswrapper[4575]: I1004 04:54:48.244633 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 04 04:54:48 crc kubenswrapper[4575]: I1004 04:54:48.307182 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="21199967-fd4f-43af-acd4-653828f2335a" containerName="galera" probeResult="failure" output=< Oct 04 04:54:48 crc kubenswrapper[4575]: wsrep_local_state_comment (Joined) differs from Synced Oct 04 04:54:48 crc kubenswrapper[4575]: > Oct 04 04:54:48 crc kubenswrapper[4575]: I1004 04:54:48.588015 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" event={"ID":"12365747-76ba-4bcb-a828-147712279d49","Type":"ContainerStarted","Data":"67dc4c51cb6b0f7f70e31d30139b282d6140c7bc3e753f41ab9db4cfd414380e"} Oct 04 04:54:48 crc kubenswrapper[4575]: I1004 04:54:48.588480 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:48 crc kubenswrapper[4575]: I1004 04:54:48.589548 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xpsnf" event={"ID":"ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3","Type":"ContainerStarted","Data":"02b30e2033e6f82d4de02a1798b299ea919883f0bbb9de0a5a566d899f40d4a1"} Oct 04 04:54:48 crc kubenswrapper[4575]: I1004 04:54:48.621548 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" podStartSLOduration=6.6215199909999995 podStartE2EDuration="6.621519991s" podCreationTimestamp="2025-10-04 04:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:48.618231276 +0000 UTC m=+1239.946790090" watchObservedRunningTime="2025-10-04 04:54:48.621519991 +0000 UTC m=+1239.950078805" Oct 04 04:54:48 crc kubenswrapper[4575]: I1004 04:54:48.646928 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-xpsnf" podStartSLOduration=3.817432475 podStartE2EDuration="7.646896271s" podCreationTimestamp="2025-10-04 04:54:41 +0000 UTC" firstStartedPulling="2025-10-04 04:54:43.052236188 +0000 UTC m=+1234.380795002" lastFinishedPulling="2025-10-04 04:54:46.881699984 +0000 UTC m=+1238.210258798" observedRunningTime="2025-10-04 04:54:48.640205489 +0000 UTC m=+1239.968764303" watchObservedRunningTime="2025-10-04 04:54:48.646896271 +0000 UTC m=+1239.975455075" Oct 04 04:54:49 crc kubenswrapper[4575]: I1004 04:54:49.301108 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-95h4j" podUID="e4e95f59-8c21-495d-bc35-151f36e37681" containerName="ovn-controller" probeResult="failure" output=< Oct 04 04:54:49 crc kubenswrapper[4575]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 04:54:49 crc kubenswrapper[4575]: > Oct 04 04:54:49 crc kubenswrapper[4575]: I1004 04:54:49.333955 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:54:49 crc kubenswrapper[4575]: I1004 04:54:49.340807 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-c66wt" Oct 04 04:54:52 crc kubenswrapper[4575]: I1004 04:54:52.552532 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 04 04:54:52 crc kubenswrapper[4575]: I1004 04:54:52.602087 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="04440b88-f2b8-418c-9676-714f0e8b7112" containerName="galera" probeResult="failure" output=< Oct 04 04:54:52 crc kubenswrapper[4575]: wsrep_local_state_comment (Joined) differs from Synced Oct 04 04:54:52 crc kubenswrapper[4575]: > Oct 04 04:54:52 crc kubenswrapper[4575]: I1004 04:54:52.617665 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bd50f51c-ef8b-4966-a7d5-3f5420295d17","Type":"ContainerStarted","Data":"7dbe845c823cd0a856feed33d671652936c2e0d01ac39adde9f2d09643aaaba8"} Oct 04 04:54:54 crc kubenswrapper[4575]: I1004 04:54:54.090732 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:54:54 crc kubenswrapper[4575]: I1004 04:54:54.110886 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=55.534515319 podStartE2EDuration="1m3.110862263s" podCreationTimestamp="2025-10-04 04:53:51 +0000 UTC" firstStartedPulling="2025-10-04 04:54:05.79918458 +0000 UTC m=+1197.127743404" lastFinishedPulling="2025-10-04 04:54:13.375531534 +0000 UTC m=+1204.704090348" observedRunningTime="2025-10-04 04:54:52.645358062 +0000 UTC m=+1243.973916876" watchObservedRunningTime="2025-10-04 04:54:54.110862263 +0000 UTC m=+1245.439421107" Oct 04 04:54:54 crc kubenswrapper[4575]: I1004 04:54:54.132784 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 04:54:54 crc kubenswrapper[4575]: I1004 04:54:54.155661 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-48z8b"] Oct 04 04:54:54 crc kubenswrapper[4575]: I1004 04:54:54.155932 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" podUID="4c489f23-9a31-437f-ae9c-3f421ff5aace" containerName="dnsmasq-dns" containerID="cri-o://5dbd852624bba2661f3547c0d8514d697c6e211a17dd4411df1ca73178bd85b9" gracePeriod=10 Oct 04 04:54:54 crc kubenswrapper[4575]: I1004 04:54:54.274657 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-95h4j" podUID="e4e95f59-8c21-495d-bc35-151f36e37681" containerName="ovn-controller" probeResult="failure" output=< Oct 04 04:54:54 crc kubenswrapper[4575]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 04:54:54 crc kubenswrapper[4575]: > Oct 04 04:54:54 crc kubenswrapper[4575]: I1004 04:54:54.495220 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:54:54 crc kubenswrapper[4575]: I1004 04:54:54.632479 4575 generic.go:334] "Generic (PLEG): container finished" podID="4c489f23-9a31-437f-ae9c-3f421ff5aace" containerID="5dbd852624bba2661f3547c0d8514d697c6e211a17dd4411df1ca73178bd85b9" exitCode=0 Oct 04 04:54:54 crc kubenswrapper[4575]: I1004 04:54:54.632829 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" event={"ID":"4c489f23-9a31-437f-ae9c-3f421ff5aace","Type":"ContainerDied","Data":"5dbd852624bba2661f3547c0d8514d697c6e211a17dd4411df1ca73178bd85b9"} Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.349086 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.395372 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px5rn\" (UniqueName: \"kubernetes.io/projected/4c489f23-9a31-437f-ae9c-3f421ff5aace-kube-api-access-px5rn\") pod \"4c489f23-9a31-437f-ae9c-3f421ff5aace\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.395460 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-dns-svc\") pod \"4c489f23-9a31-437f-ae9c-3f421ff5aace\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.395496 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-config\") pod \"4c489f23-9a31-437f-ae9c-3f421ff5aace\" (UID: \"4c489f23-9a31-437f-ae9c-3f421ff5aace\") " Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.435278 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c489f23-9a31-437f-ae9c-3f421ff5aace-kube-api-access-px5rn" (OuterVolumeSpecName: "kube-api-access-px5rn") pod "4c489f23-9a31-437f-ae9c-3f421ff5aace" (UID: "4c489f23-9a31-437f-ae9c-3f421ff5aace"). InnerVolumeSpecName "kube-api-access-px5rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.466734 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-config" (OuterVolumeSpecName: "config") pod "4c489f23-9a31-437f-ae9c-3f421ff5aace" (UID: "4c489f23-9a31-437f-ae9c-3f421ff5aace"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.481169 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c489f23-9a31-437f-ae9c-3f421ff5aace" (UID: "4c489f23-9a31-437f-ae9c-3f421ff5aace"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.497225 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.497261 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px5rn\" (UniqueName: \"kubernetes.io/projected/4c489f23-9a31-437f-ae9c-3f421ff5aace-kube-api-access-px5rn\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.497274 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c489f23-9a31-437f-ae9c-3f421ff5aace-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.640280 4575 generic.go:334] "Generic (PLEG): container finished" podID="d5724856-06a7-4b89-a67e-8239308e798f" containerID="e3c14dbe8ba059270e8e4625059100fd4118169cac6e704e71a896c5c3ab8256" exitCode=0 Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.640325 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jw946" event={"ID":"d5724856-06a7-4b89-a67e-8239308e798f","Type":"ContainerDied","Data":"e3c14dbe8ba059270e8e4625059100fd4118169cac6e704e71a896c5c3ab8256"} Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.642165 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" event={"ID":"4c489f23-9a31-437f-ae9c-3f421ff5aace","Type":"ContainerDied","Data":"6cfd0caba94e91ca91a3c44c8e72fe619050e7541bfe4179f32b63d2ca6907b2"} Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.642204 4575 scope.go:117] "RemoveContainer" containerID="5dbd852624bba2661f3547c0d8514d697c6e211a17dd4411df1ca73178bd85b9" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.642172 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-48z8b" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.644385 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ca00752a-2943-4e70-962d-91503f9dd347","Type":"ContainerStarted","Data":"85a1025fb39f733c83d20846fde736d4e1afba150025c7a1d0a5303a012a9726"} Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.662891 4575 scope.go:117] "RemoveContainer" containerID="e8f6fa0a47d77f414f6328b8b0c2a024159998ea065b75a9fdf66df5114b88a4" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.684109 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=50.569080445 podStartE2EDuration="1m0.684088633s" podCreationTimestamp="2025-10-04 04:53:55 +0000 UTC" firstStartedPulling="2025-10-04 04:54:03.124137509 +0000 UTC m=+1194.452696313" lastFinishedPulling="2025-10-04 04:54:13.239145687 +0000 UTC m=+1204.567704501" observedRunningTime="2025-10-04 04:54:55.681951392 +0000 UTC m=+1247.010510226" watchObservedRunningTime="2025-10-04 04:54:55.684088633 +0000 UTC m=+1247.012647447" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.727752 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-48z8b"] Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.734918 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-48z8b"] Oct 04 04:54:55 crc kubenswrapper[4575]: E1004 04:54:55.849974 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c489f23_9a31_437f_ae9c_3f421ff5aace.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c489f23_9a31_437f_ae9c_3f421ff5aace.slice/crio-6cfd0caba94e91ca91a3c44c8e72fe619050e7541bfe4179f32b63d2ca6907b2\": RecentStats: unable to find data in memory cache]" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.935062 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-95h4j-config-z4wxx"] Oct 04 04:54:55 crc kubenswrapper[4575]: E1004 04:54:55.935401 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a94cdf5d-d25e-4806-aec2-8a3958118c6f" containerName="init" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.935421 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a94cdf5d-d25e-4806-aec2-8a3958118c6f" containerName="init" Oct 04 04:54:55 crc kubenswrapper[4575]: E1004 04:54:55.935447 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c489f23-9a31-437f-ae9c-3f421ff5aace" containerName="dnsmasq-dns" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.935455 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c489f23-9a31-437f-ae9c-3f421ff5aace" containerName="dnsmasq-dns" Oct 04 04:54:55 crc kubenswrapper[4575]: E1004 04:54:55.935467 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c489f23-9a31-437f-ae9c-3f421ff5aace" containerName="init" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.935475 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c489f23-9a31-437f-ae9c-3f421ff5aace" containerName="init" Oct 04 04:54:55 crc kubenswrapper[4575]: E1004 04:54:55.935484 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a94cdf5d-d25e-4806-aec2-8a3958118c6f" containerName="dnsmasq-dns" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.935490 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a94cdf5d-d25e-4806-aec2-8a3958118c6f" containerName="dnsmasq-dns" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.935683 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="a94cdf5d-d25e-4806-aec2-8a3958118c6f" containerName="dnsmasq-dns" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.935698 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c489f23-9a31-437f-ae9c-3f421ff5aace" containerName="dnsmasq-dns" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.936256 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.938349 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 04:54:55 crc kubenswrapper[4575]: I1004 04:54:55.956873 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-95h4j-config-z4wxx"] Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.003912 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-scripts\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.003966 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-log-ovn\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.004039 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-additional-scripts\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.004060 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.004681 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6cvm\" (UniqueName: \"kubernetes.io/projected/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-kube-api-access-x6cvm\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.004716 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run-ovn\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.045544 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.047296 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.050312 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-xnd8t" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.050567 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.051074 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.053934 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.062236 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.105985 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jvvb\" (UniqueName: \"kubernetes.io/projected/2ee38535-11df-486c-a72d-db4a798d4969-kube-api-access-6jvvb\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.106312 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run-ovn\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.106800 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ee38535-11df-486c-a72d-db4a798d4969-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.106970 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee38535-11df-486c-a72d-db4a798d4969-config\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.107112 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-scripts\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.107188 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ee38535-11df-486c-a72d-db4a798d4969-scripts\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.107265 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-log-ovn\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.107367 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2ee38535-11df-486c-a72d-db4a798d4969-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.107457 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee38535-11df-486c-a72d-db4a798d4969-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.107567 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-additional-scripts\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.107651 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ee38535-11df-486c-a72d-db4a798d4969-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.107760 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.107879 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6cvm\" (UniqueName: \"kubernetes.io/projected/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-kube-api-access-x6cvm\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.106751 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run-ovn\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.109233 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-log-ovn\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.109870 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-additional-scripts\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.109928 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.110617 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-scripts\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.129806 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6cvm\" (UniqueName: \"kubernetes.io/projected/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-kube-api-access-x6cvm\") pod \"ovn-controller-95h4j-config-z4wxx\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.209980 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ee38535-11df-486c-a72d-db4a798d4969-scripts\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.210331 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2ee38535-11df-486c-a72d-db4a798d4969-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.210508 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee38535-11df-486c-a72d-db4a798d4969-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.210854 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2ee38535-11df-486c-a72d-db4a798d4969-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.211024 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ee38535-11df-486c-a72d-db4a798d4969-scripts\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.211132 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ee38535-11df-486c-a72d-db4a798d4969-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.211318 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jvvb\" (UniqueName: \"kubernetes.io/projected/2ee38535-11df-486c-a72d-db4a798d4969-kube-api-access-6jvvb\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.211363 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ee38535-11df-486c-a72d-db4a798d4969-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.211410 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee38535-11df-486c-a72d-db4a798d4969-config\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.212100 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee38535-11df-486c-a72d-db4a798d4969-config\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.214147 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee38535-11df-486c-a72d-db4a798d4969-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.214871 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ee38535-11df-486c-a72d-db4a798d4969-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.214974 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ee38535-11df-486c-a72d-db4a798d4969-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.228412 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jvvb\" (UniqueName: \"kubernetes.io/projected/2ee38535-11df-486c-a72d-db4a798d4969-kube-api-access-6jvvb\") pod \"ovn-northd-0\" (UID: \"2ee38535-11df-486c-a72d-db4a798d4969\") " pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.256954 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.364351 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.639714 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.889163 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-95h4j-config-z4wxx"] Oct 04 04:54:56 crc kubenswrapper[4575]: W1004 04:54:56.942429 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf15907d6_c3cf_4295_acc2_56bfba2c9fa0.slice/crio-49f2e6834fd5aeb4ec29dfb34f216ff5b3ad8d97088599f317e7ce7c7b8b7050 WatchSource:0}: Error finding container 49f2e6834fd5aeb4ec29dfb34f216ff5b3ad8d97088599f317e7ce7c7b8b7050: Status 404 returned error can't find the container with id 49f2e6834fd5aeb4ec29dfb34f216ff5b3ad8d97088599f317e7ce7c7b8b7050 Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.979387 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-z44gg"] Oct 04 04:54:56 crc kubenswrapper[4575]: I1004 04:54:56.989213 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z44gg" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.008729 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-z44gg"] Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.026752 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz9ml\" (UniqueName: \"kubernetes.io/projected/9a9115a4-74b2-49dc-90b3-d7b95bc391f4-kube-api-access-cz9ml\") pod \"keystone-db-create-z44gg\" (UID: \"9a9115a4-74b2-49dc-90b3-d7b95bc391f4\") " pod="openstack/keystone-db-create-z44gg" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.086076 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.128896 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz9ml\" (UniqueName: \"kubernetes.io/projected/9a9115a4-74b2-49dc-90b3-d7b95bc391f4-kube-api-access-cz9ml\") pod \"keystone-db-create-z44gg\" (UID: \"9a9115a4-74b2-49dc-90b3-d7b95bc391f4\") " pod="openstack/keystone-db-create-z44gg" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.152393 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.157032 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz9ml\" (UniqueName: \"kubernetes.io/projected/9a9115a4-74b2-49dc-90b3-d7b95bc391f4-kube-api-access-cz9ml\") pod \"keystone-db-create-z44gg\" (UID: \"9a9115a4-74b2-49dc-90b3-d7b95bc391f4\") " pod="openstack/keystone-db-create-z44gg" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.259404 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-g2n66"] Oct 04 04:54:57 crc kubenswrapper[4575]: E1004 04:54:57.259821 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5724856-06a7-4b89-a67e-8239308e798f" containerName="swift-ring-rebalance" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.259845 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5724856-06a7-4b89-a67e-8239308e798f" containerName="swift-ring-rebalance" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.260041 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5724856-06a7-4b89-a67e-8239308e798f" containerName="swift-ring-rebalance" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.260678 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g2n66" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.266630 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-g2n66"] Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.328742 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c489f23-9a31-437f-ae9c-3f421ff5aace" path="/var/lib/kubelet/pods/4c489f23-9a31-437f-ae9c-3f421ff5aace/volumes" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.339156 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-dispersionconf\") pod \"d5724856-06a7-4b89-a67e-8239308e798f\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.339289 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-ring-data-devices\") pod \"d5724856-06a7-4b89-a67e-8239308e798f\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.339324 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-scripts\") pod \"d5724856-06a7-4b89-a67e-8239308e798f\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.339375 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-swiftconf\") pod \"d5724856-06a7-4b89-a67e-8239308e798f\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.339409 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-combined-ca-bundle\") pod \"d5724856-06a7-4b89-a67e-8239308e798f\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.339493 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5724856-06a7-4b89-a67e-8239308e798f-etc-swift\") pod \"d5724856-06a7-4b89-a67e-8239308e798f\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.339553 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmzdb\" (UniqueName: \"kubernetes.io/projected/d5724856-06a7-4b89-a67e-8239308e798f-kube-api-access-tmzdb\") pod \"d5724856-06a7-4b89-a67e-8239308e798f\" (UID: \"d5724856-06a7-4b89-a67e-8239308e798f\") " Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.344097 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d5724856-06a7-4b89-a67e-8239308e798f" (UID: "d5724856-06a7-4b89-a67e-8239308e798f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.345367 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5724856-06a7-4b89-a67e-8239308e798f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d5724856-06a7-4b89-a67e-8239308e798f" (UID: "d5724856-06a7-4b89-a67e-8239308e798f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.350221 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5724856-06a7-4b89-a67e-8239308e798f-kube-api-access-tmzdb" (OuterVolumeSpecName: "kube-api-access-tmzdb") pod "d5724856-06a7-4b89-a67e-8239308e798f" (UID: "d5724856-06a7-4b89-a67e-8239308e798f"). InnerVolumeSpecName "kube-api-access-tmzdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.371102 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d5724856-06a7-4b89-a67e-8239308e798f" (UID: "d5724856-06a7-4b89-a67e-8239308e798f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.377877 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z44gg" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.381908 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5724856-06a7-4b89-a67e-8239308e798f" (UID: "d5724856-06a7-4b89-a67e-8239308e798f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.396843 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d5724856-06a7-4b89-a67e-8239308e798f" (UID: "d5724856-06a7-4b89-a67e-8239308e798f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.399410 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-scripts" (OuterVolumeSpecName: "scripts") pod "d5724856-06a7-4b89-a67e-8239308e798f" (UID: "d5724856-06a7-4b89-a67e-8239308e798f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.441306 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcwh5\" (UniqueName: \"kubernetes.io/projected/e23643b7-318b-4390-a9c3-346237015c04-kube-api-access-pcwh5\") pod \"placement-db-create-g2n66\" (UID: \"e23643b7-318b-4390-a9c3-346237015c04\") " pod="openstack/placement-db-create-g2n66" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.441529 4575 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.442081 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5724856-06a7-4b89-a67e-8239308e798f-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.442219 4575 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.442238 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.442248 4575 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d5724856-06a7-4b89-a67e-8239308e798f-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.442257 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmzdb\" (UniqueName: \"kubernetes.io/projected/d5724856-06a7-4b89-a67e-8239308e798f-kube-api-access-tmzdb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.442265 4575 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d5724856-06a7-4b89-a67e-8239308e798f-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.491055 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.543454 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcwh5\" (UniqueName: \"kubernetes.io/projected/e23643b7-318b-4390-a9c3-346237015c04-kube-api-access-pcwh5\") pod \"placement-db-create-g2n66\" (UID: \"e23643b7-318b-4390-a9c3-346237015c04\") " pod="openstack/placement-db-create-g2n66" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.566573 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcwh5\" (UniqueName: \"kubernetes.io/projected/e23643b7-318b-4390-a9c3-346237015c04-kube-api-access-pcwh5\") pod \"placement-db-create-g2n66\" (UID: \"e23643b7-318b-4390-a9c3-346237015c04\") " pod="openstack/placement-db-create-g2n66" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.576955 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g2n66" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.666767 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-95h4j-config-z4wxx" event={"ID":"f15907d6-c3cf-4295-acc2-56bfba2c9fa0","Type":"ContainerStarted","Data":"797b6d89affe9a8940efd0f673e8e991f4a5490d6f2839f11417625c7856ecb7"} Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.666808 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-95h4j-config-z4wxx" event={"ID":"f15907d6-c3cf-4295-acc2-56bfba2c9fa0","Type":"ContainerStarted","Data":"49f2e6834fd5aeb4ec29dfb34f216ff5b3ad8d97088599f317e7ce7c7b8b7050"} Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.670016 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jw946" event={"ID":"d5724856-06a7-4b89-a67e-8239308e798f","Type":"ContainerDied","Data":"f809bffca6509f39aa7b74c4e2c41817cb4d11a7d0367dcae01a63b554c4fdcd"} Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.670051 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f809bffca6509f39aa7b74c4e2c41817cb4d11a7d0367dcae01a63b554c4fdcd" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.670118 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jw946" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.674209 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2ee38535-11df-486c-a72d-db4a798d4969","Type":"ContainerStarted","Data":"095d172c3a8ad62eb85a5985ae07f11a43430b5ea934041abdfcff46def24786"} Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.713483 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-95h4j-config-z4wxx" podStartSLOduration=2.7134667070000003 podStartE2EDuration="2.713466707s" podCreationTimestamp="2025-10-04 04:54:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:57.712083307 +0000 UTC m=+1249.040642141" watchObservedRunningTime="2025-10-04 04:54:57.713466707 +0000 UTC m=+1249.042025521" Oct 04 04:54:57 crc kubenswrapper[4575]: I1004 04:54:57.927393 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-z44gg"] Oct 04 04:54:58 crc kubenswrapper[4575]: I1004 04:54:58.161569 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-g2n66"] Oct 04 04:54:58 crc kubenswrapper[4575]: I1004 04:54:58.708102 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g2n66" event={"ID":"e23643b7-318b-4390-a9c3-346237015c04","Type":"ContainerStarted","Data":"ee3d7346db2a34dc02cac047164fb61bb86d8f8eca1d120759a968d5f8b9d1f5"} Oct 04 04:54:58 crc kubenswrapper[4575]: I1004 04:54:58.708445 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g2n66" event={"ID":"e23643b7-318b-4390-a9c3-346237015c04","Type":"ContainerStarted","Data":"82a8bece10ff652a2ce16d5ebe4571fd3fc5ead65750771aa6c12b3f96685c8f"} Oct 04 04:54:58 crc kubenswrapper[4575]: I1004 04:54:58.709877 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z44gg" event={"ID":"9a9115a4-74b2-49dc-90b3-d7b95bc391f4","Type":"ContainerStarted","Data":"c0993731dbece24a7e8c805c5361db9d7f515ca46b84e427925da9672422c0e3"} Oct 04 04:54:58 crc kubenswrapper[4575]: I1004 04:54:58.709895 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z44gg" event={"ID":"9a9115a4-74b2-49dc-90b3-d7b95bc391f4","Type":"ContainerStarted","Data":"e9f73d0825238e9bd6a202031b3d0e62c0cef81ae1528953fb666f42d13e9ccd"} Oct 04 04:54:58 crc kubenswrapper[4575]: I1004 04:54:58.713425 4575 generic.go:334] "Generic (PLEG): container finished" podID="f15907d6-c3cf-4295-acc2-56bfba2c9fa0" containerID="797b6d89affe9a8940efd0f673e8e991f4a5490d6f2839f11417625c7856ecb7" exitCode=0 Oct 04 04:54:58 crc kubenswrapper[4575]: I1004 04:54:58.713505 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-95h4j-config-z4wxx" event={"ID":"f15907d6-c3cf-4295-acc2-56bfba2c9fa0","Type":"ContainerDied","Data":"797b6d89affe9a8940efd0f673e8e991f4a5490d6f2839f11417625c7856ecb7"} Oct 04 04:54:58 crc kubenswrapper[4575]: I1004 04:54:58.732953 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-z44gg" podStartSLOduration=2.732933005 podStartE2EDuration="2.732933005s" podCreationTimestamp="2025-10-04 04:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:58.723521854 +0000 UTC m=+1250.052080668" watchObservedRunningTime="2025-10-04 04:54:58.732933005 +0000 UTC m=+1250.061491819" Oct 04 04:54:59 crc kubenswrapper[4575]: I1004 04:54:59.229389 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-95h4j" Oct 04 04:54:59 crc kubenswrapper[4575]: I1004 04:54:59.744404 4575 generic.go:334] "Generic (PLEG): container finished" podID="9a9115a4-74b2-49dc-90b3-d7b95bc391f4" containerID="c0993731dbece24a7e8c805c5361db9d7f515ca46b84e427925da9672422c0e3" exitCode=0 Oct 04 04:54:59 crc kubenswrapper[4575]: I1004 04:54:59.745757 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z44gg" event={"ID":"9a9115a4-74b2-49dc-90b3-d7b95bc391f4","Type":"ContainerDied","Data":"c0993731dbece24a7e8c805c5361db9d7f515ca46b84e427925da9672422c0e3"} Oct 04 04:54:59 crc kubenswrapper[4575]: I1004 04:54:59.769714 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-g2n66" podStartSLOduration=2.769692682 podStartE2EDuration="2.769692682s" podCreationTimestamp="2025-10-04 04:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:59.762806464 +0000 UTC m=+1251.091365288" watchObservedRunningTime="2025-10-04 04:54:59.769692682 +0000 UTC m=+1251.098251496" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.040798 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.190611 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-additional-scripts\") pod \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.190678 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6cvm\" (UniqueName: \"kubernetes.io/projected/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-kube-api-access-x6cvm\") pod \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.190725 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-scripts\") pod \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.190773 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run-ovn\") pod \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.190828 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-log-ovn\") pod \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.190872 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run\") pod \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\" (UID: \"f15907d6-c3cf-4295-acc2-56bfba2c9fa0\") " Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.191359 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run" (OuterVolumeSpecName: "var-run") pod "f15907d6-c3cf-4295-acc2-56bfba2c9fa0" (UID: "f15907d6-c3cf-4295-acc2-56bfba2c9fa0"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.191424 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f15907d6-c3cf-4295-acc2-56bfba2c9fa0" (UID: "f15907d6-c3cf-4295-acc2-56bfba2c9fa0"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.191442 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f15907d6-c3cf-4295-acc2-56bfba2c9fa0" (UID: "f15907d6-c3cf-4295-acc2-56bfba2c9fa0"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.191976 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f15907d6-c3cf-4295-acc2-56bfba2c9fa0" (UID: "f15907d6-c3cf-4295-acc2-56bfba2c9fa0"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.192358 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-scripts" (OuterVolumeSpecName: "scripts") pod "f15907d6-c3cf-4295-acc2-56bfba2c9fa0" (UID: "f15907d6-c3cf-4295-acc2-56bfba2c9fa0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.199860 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-kube-api-access-x6cvm" (OuterVolumeSpecName: "kube-api-access-x6cvm") pod "f15907d6-c3cf-4295-acc2-56bfba2c9fa0" (UID: "f15907d6-c3cf-4295-acc2-56bfba2c9fa0"). InnerVolumeSpecName "kube-api-access-x6cvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.293243 4575 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.293777 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6cvm\" (UniqueName: \"kubernetes.io/projected/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-kube-api-access-x6cvm\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.293882 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.293961 4575 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.294039 4575 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.294113 4575 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f15907d6-c3cf-4295-acc2-56bfba2c9fa0-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.753528 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-95h4j-config-z4wxx" event={"ID":"f15907d6-c3cf-4295-acc2-56bfba2c9fa0","Type":"ContainerDied","Data":"49f2e6834fd5aeb4ec29dfb34f216ff5b3ad8d97088599f317e7ce7c7b8b7050"} Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.753579 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j-config-z4wxx" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.753613 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49f2e6834fd5aeb4ec29dfb34f216ff5b3ad8d97088599f317e7ce7c7b8b7050" Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.755244 4575 generic.go:334] "Generic (PLEG): container finished" podID="e23643b7-318b-4390-a9c3-346237015c04" containerID="ee3d7346db2a34dc02cac047164fb61bb86d8f8eca1d120759a968d5f8b9d1f5" exitCode=0 Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.755402 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g2n66" event={"ID":"e23643b7-318b-4390-a9c3-346237015c04","Type":"ContainerDied","Data":"ee3d7346db2a34dc02cac047164fb61bb86d8f8eca1d120759a968d5f8b9d1f5"} Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.819624 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-95h4j-config-z4wxx"] Oct 04 04:55:00 crc kubenswrapper[4575]: I1004 04:55:00.833188 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-95h4j-config-z4wxx"] Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.048278 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-95h4j-config-ppz7v"] Oct 04 04:55:01 crc kubenswrapper[4575]: E1004 04:55:01.048906 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f15907d6-c3cf-4295-acc2-56bfba2c9fa0" containerName="ovn-config" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.048919 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f15907d6-c3cf-4295-acc2-56bfba2c9fa0" containerName="ovn-config" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.049071 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f15907d6-c3cf-4295-acc2-56bfba2c9fa0" containerName="ovn-config" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.049558 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.060728 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.106455 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-95h4j-config-ppz7v"] Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.220236 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-additional-scripts\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.220283 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-scripts\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.220346 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.220428 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qpln\" (UniqueName: \"kubernetes.io/projected/b0f6e010-4289-4541-941c-af961b45a613-kube-api-access-6qpln\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.220473 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-log-ovn\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.220500 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run-ovn\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.236884 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z44gg" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.325162 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz9ml\" (UniqueName: \"kubernetes.io/projected/9a9115a4-74b2-49dc-90b3-d7b95bc391f4-kube-api-access-cz9ml\") pod \"9a9115a4-74b2-49dc-90b3-d7b95bc391f4\" (UID: \"9a9115a4-74b2-49dc-90b3-d7b95bc391f4\") " Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.325711 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-additional-scripts\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.325804 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-scripts\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.325908 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.326009 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qpln\" (UniqueName: \"kubernetes.io/projected/b0f6e010-4289-4541-941c-af961b45a613-kube-api-access-6qpln\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.326098 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-log-ovn\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.326171 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run-ovn\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.326540 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run-ovn\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.327385 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-additional-scripts\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.328866 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-scripts\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.329024 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-log-ovn\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.329076 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.329648 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f15907d6-c3cf-4295-acc2-56bfba2c9fa0" path="/var/lib/kubelet/pods/f15907d6-c3cf-4295-acc2-56bfba2c9fa0/volumes" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.337893 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a9115a4-74b2-49dc-90b3-d7b95bc391f4-kube-api-access-cz9ml" (OuterVolumeSpecName: "kube-api-access-cz9ml") pod "9a9115a4-74b2-49dc-90b3-d7b95bc391f4" (UID: "9a9115a4-74b2-49dc-90b3-d7b95bc391f4"). InnerVolumeSpecName "kube-api-access-cz9ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.352556 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qpln\" (UniqueName: \"kubernetes.io/projected/b0f6e010-4289-4541-941c-af961b45a613-kube-api-access-6qpln\") pod \"ovn-controller-95h4j-config-ppz7v\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.427644 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz9ml\" (UniqueName: \"kubernetes.io/projected/9a9115a4-74b2-49dc-90b3-d7b95bc391f4-kube-api-access-cz9ml\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.528832 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.775956 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2ee38535-11df-486c-a72d-db4a798d4969","Type":"ContainerStarted","Data":"cf639dc514d9fe5fc28ad3a1d7d1db189a33855ada3e9a3e167837d79a3feee1"} Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.776251 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.776262 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2ee38535-11df-486c-a72d-db4a798d4969","Type":"ContainerStarted","Data":"66484237fa35940a8279c28ece1335d8de3ee48c11cd6ea77ac7ac4f9aee02a9"} Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.779116 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z44gg" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.780103 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z44gg" event={"ID":"9a9115a4-74b2-49dc-90b3-d7b95bc391f4","Type":"ContainerDied","Data":"e9f73d0825238e9bd6a202031b3d0e62c0cef81ae1528953fb666f42d13e9ccd"} Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.780136 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9f73d0825238e9bd6a202031b3d0e62c0cef81ae1528953fb666f42d13e9ccd" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.806691 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.32808887 podStartE2EDuration="5.806668723s" podCreationTimestamp="2025-10-04 04:54:56 +0000 UTC" firstStartedPulling="2025-10-04 04:54:57.097557276 +0000 UTC m=+1248.426116090" lastFinishedPulling="2025-10-04 04:55:00.576137129 +0000 UTC m=+1251.904695943" observedRunningTime="2025-10-04 04:55:01.801790773 +0000 UTC m=+1253.130349587" watchObservedRunningTime="2025-10-04 04:55:01.806668723 +0000 UTC m=+1253.135227527" Oct 04 04:55:01 crc kubenswrapper[4575]: I1004 04:55:01.991907 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-95h4j-config-ppz7v"] Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.152062 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g2n66" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.240444 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcwh5\" (UniqueName: \"kubernetes.io/projected/e23643b7-318b-4390-a9c3-346237015c04-kube-api-access-pcwh5\") pod \"e23643b7-318b-4390-a9c3-346237015c04\" (UID: \"e23643b7-318b-4390-a9c3-346237015c04\") " Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.250119 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e23643b7-318b-4390-a9c3-346237015c04-kube-api-access-pcwh5" (OuterVolumeSpecName: "kube-api-access-pcwh5") pod "e23643b7-318b-4390-a9c3-346237015c04" (UID: "e23643b7-318b-4390-a9c3-346237015c04"). InnerVolumeSpecName "kube-api-access-pcwh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.342924 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcwh5\" (UniqueName: \"kubernetes.io/projected/e23643b7-318b-4390-a9c3-346237015c04-kube-api-access-pcwh5\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.556653 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-5v4gf"] Oct 04 04:55:02 crc kubenswrapper[4575]: E1004 04:55:02.557024 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a9115a4-74b2-49dc-90b3-d7b95bc391f4" containerName="mariadb-database-create" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.557039 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a9115a4-74b2-49dc-90b3-d7b95bc391f4" containerName="mariadb-database-create" Oct 04 04:55:02 crc kubenswrapper[4575]: E1004 04:55:02.557067 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23643b7-318b-4390-a9c3-346237015c04" containerName="mariadb-database-create" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.557073 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23643b7-318b-4390-a9c3-346237015c04" containerName="mariadb-database-create" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.557217 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23643b7-318b-4390-a9c3-346237015c04" containerName="mariadb-database-create" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.557243 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a9115a4-74b2-49dc-90b3-d7b95bc391f4" containerName="mariadb-database-create" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.557810 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5v4gf" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.576606 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5v4gf"] Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.648688 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrfj8\" (UniqueName: \"kubernetes.io/projected/41d086f3-2b5e-45e5-b580-71abdd6acd71-kube-api-access-mrfj8\") pod \"glance-db-create-5v4gf\" (UID: \"41d086f3-2b5e-45e5-b580-71abdd6acd71\") " pod="openstack/glance-db-create-5v4gf" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.750876 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrfj8\" (UniqueName: \"kubernetes.io/projected/41d086f3-2b5e-45e5-b580-71abdd6acd71-kube-api-access-mrfj8\") pod \"glance-db-create-5v4gf\" (UID: \"41d086f3-2b5e-45e5-b580-71abdd6acd71\") " pod="openstack/glance-db-create-5v4gf" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.771167 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrfj8\" (UniqueName: \"kubernetes.io/projected/41d086f3-2b5e-45e5-b580-71abdd6acd71-kube-api-access-mrfj8\") pod \"glance-db-create-5v4gf\" (UID: \"41d086f3-2b5e-45e5-b580-71abdd6acd71\") " pod="openstack/glance-db-create-5v4gf" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.790022 4575 generic.go:334] "Generic (PLEG): container finished" podID="b0f6e010-4289-4541-941c-af961b45a613" containerID="c1934941b3c46049e7df791739c6b54a1c9904ab83f51bff37992df55a51f8c6" exitCode=0 Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.790162 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-95h4j-config-ppz7v" event={"ID":"b0f6e010-4289-4541-941c-af961b45a613","Type":"ContainerDied","Data":"c1934941b3c46049e7df791739c6b54a1c9904ab83f51bff37992df55a51f8c6"} Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.790195 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-95h4j-config-ppz7v" event={"ID":"b0f6e010-4289-4541-941c-af961b45a613","Type":"ContainerStarted","Data":"fb2f5322ef55869dd136860e59bf6215e9a00b7b4942482cae2216f1cee7d76b"} Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.793226 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g2n66" event={"ID":"e23643b7-318b-4390-a9c3-346237015c04","Type":"ContainerDied","Data":"82a8bece10ff652a2ce16d5ebe4571fd3fc5ead65750771aa6c12b3f96685c8f"} Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.793266 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g2n66" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.793284 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82a8bece10ff652a2ce16d5ebe4571fd3fc5ead65750771aa6c12b3f96685c8f" Oct 04 04:55:02 crc kubenswrapper[4575]: I1004 04:55:02.880978 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5v4gf" Oct 04 04:55:03 crc kubenswrapper[4575]: I1004 04:55:03.258923 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:55:03 crc kubenswrapper[4575]: I1004 04:55:03.267215 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9ec5625-8b15-42bd-8258-4960d8000469-etc-swift\") pod \"swift-storage-0\" (UID: \"c9ec5625-8b15-42bd-8258-4960d8000469\") " pod="openstack/swift-storage-0" Oct 04 04:55:03 crc kubenswrapper[4575]: I1004 04:55:03.367025 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5v4gf"] Oct 04 04:55:03 crc kubenswrapper[4575]: I1004 04:55:03.466269 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 04:55:03 crc kubenswrapper[4575]: I1004 04:55:03.803666 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5v4gf" event={"ID":"41d086f3-2b5e-45e5-b580-71abdd6acd71","Type":"ContainerDied","Data":"1d36a2c8de90600d8c1f05336ae56eae60010bb2cb80c149a9b2620f21192621"} Oct 04 04:55:03 crc kubenswrapper[4575]: I1004 04:55:03.803901 4575 generic.go:334] "Generic (PLEG): container finished" podID="41d086f3-2b5e-45e5-b580-71abdd6acd71" containerID="1d36a2c8de90600d8c1f05336ae56eae60010bb2cb80c149a9b2620f21192621" exitCode=0 Oct 04 04:55:03 crc kubenswrapper[4575]: I1004 04:55:03.804288 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5v4gf" event={"ID":"41d086f3-2b5e-45e5-b580-71abdd6acd71","Type":"ContainerStarted","Data":"7c44c3905ea3136c43ef101a292a0fed4188348b27eeb56d9df222fa5cc5d1e5"} Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.112775 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 04:55:04 crc kubenswrapper[4575]: W1004 04:55:04.115540 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9ec5625_8b15_42bd_8258_4960d8000469.slice/crio-d2eac02f848659f5ce703f0c3fd2455a4577e3608fd505781a703a91ba595577 WatchSource:0}: Error finding container d2eac02f848659f5ce703f0c3fd2455a4577e3608fd505781a703a91ba595577: Status 404 returned error can't find the container with id d2eac02f848659f5ce703f0c3fd2455a4577e3608fd505781a703a91ba595577 Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.157973 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.276770 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run-ovn\") pod \"b0f6e010-4289-4541-941c-af961b45a613\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.276883 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-log-ovn\") pod \"b0f6e010-4289-4541-941c-af961b45a613\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.276954 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qpln\" (UniqueName: \"kubernetes.io/projected/b0f6e010-4289-4541-941c-af961b45a613-kube-api-access-6qpln\") pod \"b0f6e010-4289-4541-941c-af961b45a613\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.276986 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run\") pod \"b0f6e010-4289-4541-941c-af961b45a613\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.277016 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-scripts\") pod \"b0f6e010-4289-4541-941c-af961b45a613\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.277037 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-additional-scripts\") pod \"b0f6e010-4289-4541-941c-af961b45a613\" (UID: \"b0f6e010-4289-4541-941c-af961b45a613\") " Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.277865 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b0f6e010-4289-4541-941c-af961b45a613" (UID: "b0f6e010-4289-4541-941c-af961b45a613"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.277938 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b0f6e010-4289-4541-941c-af961b45a613" (UID: "b0f6e010-4289-4541-941c-af961b45a613"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.278225 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b0f6e010-4289-4541-941c-af961b45a613" (UID: "b0f6e010-4289-4541-941c-af961b45a613"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.278276 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run" (OuterVolumeSpecName: "var-run") pod "b0f6e010-4289-4541-941c-af961b45a613" (UID: "b0f6e010-4289-4541-941c-af961b45a613"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.278881 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-scripts" (OuterVolumeSpecName: "scripts") pod "b0f6e010-4289-4541-941c-af961b45a613" (UID: "b0f6e010-4289-4541-941c-af961b45a613"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.283849 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f6e010-4289-4541-941c-af961b45a613-kube-api-access-6qpln" (OuterVolumeSpecName: "kube-api-access-6qpln") pod "b0f6e010-4289-4541-941c-af961b45a613" (UID: "b0f6e010-4289-4541-941c-af961b45a613"). InnerVolumeSpecName "kube-api-access-6qpln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.379257 4575 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.379521 4575 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.379581 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qpln\" (UniqueName: \"kubernetes.io/projected/b0f6e010-4289-4541-941c-af961b45a613-kube-api-access-6qpln\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.379653 4575 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b0f6e010-4289-4541-941c-af961b45a613-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.379738 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.379813 4575 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f6e010-4289-4541-941c-af961b45a613-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.584389 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-lprm4"] Oct 04 04:55:04 crc kubenswrapper[4575]: E1004 04:55:04.584729 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f6e010-4289-4541-941c-af961b45a613" containerName="ovn-config" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.584746 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f6e010-4289-4541-941c-af961b45a613" containerName="ovn-config" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.584972 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0f6e010-4289-4541-941c-af961b45a613" containerName="ovn-config" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.585616 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lprm4" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.594670 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lprm4"] Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.684863 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkskx\" (UniqueName: \"kubernetes.io/projected/c5bb6f33-61f5-4b6f-8449-2595b8593d28-kube-api-access-lkskx\") pod \"cinder-db-create-lprm4\" (UID: \"c5bb6f33-61f5-4b6f-8449-2595b8593d28\") " pod="openstack/cinder-db-create-lprm4" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.777177 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-72mrw"] Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.778336 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-72mrw" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.786989 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkskx\" (UniqueName: \"kubernetes.io/projected/c5bb6f33-61f5-4b6f-8449-2595b8593d28-kube-api-access-lkskx\") pod \"cinder-db-create-lprm4\" (UID: \"c5bb6f33-61f5-4b6f-8449-2595b8593d28\") " pod="openstack/cinder-db-create-lprm4" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.790637 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-72mrw"] Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.809488 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkskx\" (UniqueName: \"kubernetes.io/projected/c5bb6f33-61f5-4b6f-8449-2595b8593d28-kube-api-access-lkskx\") pod \"cinder-db-create-lprm4\" (UID: \"c5bb6f33-61f5-4b6f-8449-2595b8593d28\") " pod="openstack/cinder-db-create-lprm4" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.817829 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-95h4j-config-ppz7v" event={"ID":"b0f6e010-4289-4541-941c-af961b45a613","Type":"ContainerDied","Data":"fb2f5322ef55869dd136860e59bf6215e9a00b7b4942482cae2216f1cee7d76b"} Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.818262 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb2f5322ef55869dd136860e59bf6215e9a00b7b4942482cae2216f1cee7d76b" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.818133 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-95h4j-config-ppz7v" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.819032 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"d2eac02f848659f5ce703f0c3fd2455a4577e3608fd505781a703a91ba595577"} Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.888915 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k77th\" (UniqueName: \"kubernetes.io/projected/eea4fe86-b04b-44d4-bc68-361b6bf34b6a-kube-api-access-k77th\") pod \"barbican-db-create-72mrw\" (UID: \"eea4fe86-b04b-44d4-bc68-361b6bf34b6a\") " pod="openstack/barbican-db-create-72mrw" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.909879 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lprm4" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.910710 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-kjpkv"] Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.912979 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kjpkv" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.926969 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-kjpkv"] Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.991109 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k77th\" (UniqueName: \"kubernetes.io/projected/eea4fe86-b04b-44d4-bc68-361b6bf34b6a-kube-api-access-k77th\") pod \"barbican-db-create-72mrw\" (UID: \"eea4fe86-b04b-44d4-bc68-361b6bf34b6a\") " pod="openstack/barbican-db-create-72mrw" Oct 04 04:55:04 crc kubenswrapper[4575]: I1004 04:55:04.991241 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxxhk\" (UniqueName: \"kubernetes.io/projected/bedd9a1e-02e4-4fac-8f24-f45234f3aa40-kube-api-access-fxxhk\") pod \"neutron-db-create-kjpkv\" (UID: \"bedd9a1e-02e4-4fac-8f24-f45234f3aa40\") " pod="openstack/neutron-db-create-kjpkv" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.014764 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k77th\" (UniqueName: \"kubernetes.io/projected/eea4fe86-b04b-44d4-bc68-361b6bf34b6a-kube-api-access-k77th\") pod \"barbican-db-create-72mrw\" (UID: \"eea4fe86-b04b-44d4-bc68-361b6bf34b6a\") " pod="openstack/barbican-db-create-72mrw" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.094683 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxxhk\" (UniqueName: \"kubernetes.io/projected/bedd9a1e-02e4-4fac-8f24-f45234f3aa40-kube-api-access-fxxhk\") pod \"neutron-db-create-kjpkv\" (UID: \"bedd9a1e-02e4-4fac-8f24-f45234f3aa40\") " pod="openstack/neutron-db-create-kjpkv" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.096791 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-72mrw" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.116739 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxxhk\" (UniqueName: \"kubernetes.io/projected/bedd9a1e-02e4-4fac-8f24-f45234f3aa40-kube-api-access-fxxhk\") pod \"neutron-db-create-kjpkv\" (UID: \"bedd9a1e-02e4-4fac-8f24-f45234f3aa40\") " pod="openstack/neutron-db-create-kjpkv" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.215813 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5v4gf" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.237087 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kjpkv" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.265669 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-95h4j-config-ppz7v"] Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.272706 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-95h4j-config-ppz7v"] Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.298356 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrfj8\" (UniqueName: \"kubernetes.io/projected/41d086f3-2b5e-45e5-b580-71abdd6acd71-kube-api-access-mrfj8\") pod \"41d086f3-2b5e-45e5-b580-71abdd6acd71\" (UID: \"41d086f3-2b5e-45e5-b580-71abdd6acd71\") " Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.304335 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41d086f3-2b5e-45e5-b580-71abdd6acd71-kube-api-access-mrfj8" (OuterVolumeSpecName: "kube-api-access-mrfj8") pod "41d086f3-2b5e-45e5-b580-71abdd6acd71" (UID: "41d086f3-2b5e-45e5-b580-71abdd6acd71"). InnerVolumeSpecName "kube-api-access-mrfj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.329150 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0f6e010-4289-4541-941c-af961b45a613" path="/var/lib/kubelet/pods/b0f6e010-4289-4541-941c-af961b45a613/volumes" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.401935 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrfj8\" (UniqueName: \"kubernetes.io/projected/41d086f3-2b5e-45e5-b580-71abdd6acd71-kube-api-access-mrfj8\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.473364 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lprm4"] Oct 04 04:55:05 crc kubenswrapper[4575]: W1004 04:55:05.476407 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5bb6f33_61f5_4b6f_8449_2595b8593d28.slice/crio-b1b30077cd379d3d20e30c52b561b9b793a0c3fce3568ca7a61913b50bd49b39 WatchSource:0}: Error finding container b1b30077cd379d3d20e30c52b561b9b793a0c3fce3568ca7a61913b50bd49b39: Status 404 returned error can't find the container with id b1b30077cd379d3d20e30c52b561b9b793a0c3fce3568ca7a61913b50bd49b39 Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.684917 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-72mrw"] Oct 04 04:55:05 crc kubenswrapper[4575]: W1004 04:55:05.690475 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeea4fe86_b04b_44d4_bc68_361b6bf34b6a.slice/crio-4e3263b172d403826622ad66f9a96b8978ff659e00ba02e88da794e81bf19ce6 WatchSource:0}: Error finding container 4e3263b172d403826622ad66f9a96b8978ff659e00ba02e88da794e81bf19ce6: Status 404 returned error can't find the container with id 4e3263b172d403826622ad66f9a96b8978ff659e00ba02e88da794e81bf19ce6 Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.809465 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-kjpkv"] Oct 04 04:55:05 crc kubenswrapper[4575]: W1004 04:55:05.811379 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbedd9a1e_02e4_4fac_8f24_f45234f3aa40.slice/crio-2e8f9faa5adc35c503e37f3bf873ce1c2bcab09fb2fcb4d2e8e8638f6fdd1274 WatchSource:0}: Error finding container 2e8f9faa5adc35c503e37f3bf873ce1c2bcab09fb2fcb4d2e8e8638f6fdd1274: Status 404 returned error can't find the container with id 2e8f9faa5adc35c503e37f3bf873ce1c2bcab09fb2fcb4d2e8e8638f6fdd1274 Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.830004 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lprm4" event={"ID":"c5bb6f33-61f5-4b6f-8449-2595b8593d28","Type":"ContainerStarted","Data":"2e207ce1c97a78cf8c52688d7420ddfaa6a5cc1fddd41f497ee1cd16f818651f"} Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.830055 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lprm4" event={"ID":"c5bb6f33-61f5-4b6f-8449-2595b8593d28","Type":"ContainerStarted","Data":"b1b30077cd379d3d20e30c52b561b9b793a0c3fce3568ca7a61913b50bd49b39"} Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.832243 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-72mrw" event={"ID":"eea4fe86-b04b-44d4-bc68-361b6bf34b6a","Type":"ContainerStarted","Data":"4e3263b172d403826622ad66f9a96b8978ff659e00ba02e88da794e81bf19ce6"} Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.833298 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kjpkv" event={"ID":"bedd9a1e-02e4-4fac-8f24-f45234f3aa40","Type":"ContainerStarted","Data":"2e8f9faa5adc35c503e37f3bf873ce1c2bcab09fb2fcb4d2e8e8638f6fdd1274"} Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.834831 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5v4gf" event={"ID":"41d086f3-2b5e-45e5-b580-71abdd6acd71","Type":"ContainerDied","Data":"7c44c3905ea3136c43ef101a292a0fed4188348b27eeb56d9df222fa5cc5d1e5"} Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.834859 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c44c3905ea3136c43ef101a292a0fed4188348b27eeb56d9df222fa5cc5d1e5" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.834907 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5v4gf" Oct 04 04:55:05 crc kubenswrapper[4575]: I1004 04:55:05.845451 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-lprm4" podStartSLOduration=1.845428644 podStartE2EDuration="1.845428644s" podCreationTimestamp="2025-10-04 04:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:05.842970384 +0000 UTC m=+1257.171529208" watchObservedRunningTime="2025-10-04 04:55:05.845428644 +0000 UTC m=+1257.173987478" Oct 04 04:55:06 crc kubenswrapper[4575]: I1004 04:55:06.843198 4575 generic.go:334] "Generic (PLEG): container finished" podID="eea4fe86-b04b-44d4-bc68-361b6bf34b6a" containerID="2c7b84617bc07cf1fca4dfa5442e1ceaeb8f429e0385dc35a6fb980d816b2835" exitCode=0 Oct 04 04:55:06 crc kubenswrapper[4575]: I1004 04:55:06.843316 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-72mrw" event={"ID":"eea4fe86-b04b-44d4-bc68-361b6bf34b6a","Type":"ContainerDied","Data":"2c7b84617bc07cf1fca4dfa5442e1ceaeb8f429e0385dc35a6fb980d816b2835"} Oct 04 04:55:06 crc kubenswrapper[4575]: I1004 04:55:06.847134 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kjpkv" event={"ID":"bedd9a1e-02e4-4fac-8f24-f45234f3aa40","Type":"ContainerStarted","Data":"51d26e723c7fcb9b4ac5cad529a9a6eb65907640b8b6128aa772a959028be65d"} Oct 04 04:55:06 crc kubenswrapper[4575]: I1004 04:55:06.849313 4575 generic.go:334] "Generic (PLEG): container finished" podID="c5bb6f33-61f5-4b6f-8449-2595b8593d28" containerID="2e207ce1c97a78cf8c52688d7420ddfaa6a5cc1fddd41f497ee1cd16f818651f" exitCode=0 Oct 04 04:55:06 crc kubenswrapper[4575]: I1004 04:55:06.849346 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lprm4" event={"ID":"c5bb6f33-61f5-4b6f-8449-2595b8593d28","Type":"ContainerDied","Data":"2e207ce1c97a78cf8c52688d7420ddfaa6a5cc1fddd41f497ee1cd16f818651f"} Oct 04 04:55:06 crc kubenswrapper[4575]: I1004 04:55:06.890181 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-kjpkv" podStartSLOduration=2.890164581 podStartE2EDuration="2.890164581s" podCreationTimestamp="2025-10-04 04:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:06.884618712 +0000 UTC m=+1258.213177526" watchObservedRunningTime="2025-10-04 04:55:06.890164581 +0000 UTC m=+1258.218723395" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.002119 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7c2a-account-create-qvx8q"] Oct 04 04:55:07 crc kubenswrapper[4575]: E1004 04:55:07.002522 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d086f3-2b5e-45e5-b580-71abdd6acd71" containerName="mariadb-database-create" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.002546 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d086f3-2b5e-45e5-b580-71abdd6acd71" containerName="mariadb-database-create" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.002775 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d086f3-2b5e-45e5-b580-71abdd6acd71" containerName="mariadb-database-create" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.003442 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c2a-account-create-qvx8q" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.005711 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.014208 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7c2a-account-create-qvx8q"] Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.038835 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mkxk\" (UniqueName: \"kubernetes.io/projected/8933fd6b-ec1a-4c42-86a2-f09d447827b6-kube-api-access-9mkxk\") pod \"keystone-7c2a-account-create-qvx8q\" (UID: \"8933fd6b-ec1a-4c42-86a2-f09d447827b6\") " pod="openstack/keystone-7c2a-account-create-qvx8q" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.140241 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mkxk\" (UniqueName: \"kubernetes.io/projected/8933fd6b-ec1a-4c42-86a2-f09d447827b6-kube-api-access-9mkxk\") pod \"keystone-7c2a-account-create-qvx8q\" (UID: \"8933fd6b-ec1a-4c42-86a2-f09d447827b6\") " pod="openstack/keystone-7c2a-account-create-qvx8q" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.167672 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mkxk\" (UniqueName: \"kubernetes.io/projected/8933fd6b-ec1a-4c42-86a2-f09d447827b6-kube-api-access-9mkxk\") pod \"keystone-7c2a-account-create-qvx8q\" (UID: \"8933fd6b-ec1a-4c42-86a2-f09d447827b6\") " pod="openstack/keystone-7c2a-account-create-qvx8q" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.322151 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c2a-account-create-qvx8q" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.389185 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-0102-account-create-p7v69"] Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.390185 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0102-account-create-p7v69" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.395876 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.399501 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0102-account-create-p7v69"] Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.444995 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbjbh\" (UniqueName: \"kubernetes.io/projected/34ce4329-745c-4427-8b49-e538401c3226-kube-api-access-lbjbh\") pod \"placement-0102-account-create-p7v69\" (UID: \"34ce4329-745c-4427-8b49-e538401c3226\") " pod="openstack/placement-0102-account-create-p7v69" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.547620 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbjbh\" (UniqueName: \"kubernetes.io/projected/34ce4329-745c-4427-8b49-e538401c3226-kube-api-access-lbjbh\") pod \"placement-0102-account-create-p7v69\" (UID: \"34ce4329-745c-4427-8b49-e538401c3226\") " pod="openstack/placement-0102-account-create-p7v69" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.581657 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbjbh\" (UniqueName: \"kubernetes.io/projected/34ce4329-745c-4427-8b49-e538401c3226-kube-api-access-lbjbh\") pod \"placement-0102-account-create-p7v69\" (UID: \"34ce4329-745c-4427-8b49-e538401c3226\") " pod="openstack/placement-0102-account-create-p7v69" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.765743 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0102-account-create-p7v69" Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.858418 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7c2a-account-create-qvx8q"] Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.864602 4575 generic.go:334] "Generic (PLEG): container finished" podID="bedd9a1e-02e4-4fac-8f24-f45234f3aa40" containerID="51d26e723c7fcb9b4ac5cad529a9a6eb65907640b8b6128aa772a959028be65d" exitCode=0 Oct 04 04:55:07 crc kubenswrapper[4575]: I1004 04:55:07.865142 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kjpkv" event={"ID":"bedd9a1e-02e4-4fac-8f24-f45234f3aa40","Type":"ContainerDied","Data":"51d26e723c7fcb9b4ac5cad529a9a6eb65907640b8b6128aa772a959028be65d"} Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.524620 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-72mrw" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.549461 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lprm4" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.562484 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k77th\" (UniqueName: \"kubernetes.io/projected/eea4fe86-b04b-44d4-bc68-361b6bf34b6a-kube-api-access-k77th\") pod \"eea4fe86-b04b-44d4-bc68-361b6bf34b6a\" (UID: \"eea4fe86-b04b-44d4-bc68-361b6bf34b6a\") " Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.577118 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eea4fe86-b04b-44d4-bc68-361b6bf34b6a-kube-api-access-k77th" (OuterVolumeSpecName: "kube-api-access-k77th") pod "eea4fe86-b04b-44d4-bc68-361b6bf34b6a" (UID: "eea4fe86-b04b-44d4-bc68-361b6bf34b6a"). InnerVolumeSpecName "kube-api-access-k77th". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.664210 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkskx\" (UniqueName: \"kubernetes.io/projected/c5bb6f33-61f5-4b6f-8449-2595b8593d28-kube-api-access-lkskx\") pod \"c5bb6f33-61f5-4b6f-8449-2595b8593d28\" (UID: \"c5bb6f33-61f5-4b6f-8449-2595b8593d28\") " Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.664532 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k77th\" (UniqueName: \"kubernetes.io/projected/eea4fe86-b04b-44d4-bc68-361b6bf34b6a-kube-api-access-k77th\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.667788 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5bb6f33-61f5-4b6f-8449-2595b8593d28-kube-api-access-lkskx" (OuterVolumeSpecName: "kube-api-access-lkskx") pod "c5bb6f33-61f5-4b6f-8449-2595b8593d28" (UID: "c5bb6f33-61f5-4b6f-8449-2595b8593d28"). InnerVolumeSpecName "kube-api-access-lkskx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.767020 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkskx\" (UniqueName: \"kubernetes.io/projected/c5bb6f33-61f5-4b6f-8449-2595b8593d28-kube-api-access-lkskx\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.857970 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0102-account-create-p7v69"] Oct 04 04:55:08 crc kubenswrapper[4575]: W1004 04:55:08.858292 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34ce4329_745c_4427_8b49_e538401c3226.slice/crio-4bddebc316e83fb9c8e8cc385955f8fb6add6e032f337b97ef3aef91e21944ac WatchSource:0}: Error finding container 4bddebc316e83fb9c8e8cc385955f8fb6add6e032f337b97ef3aef91e21944ac: Status 404 returned error can't find the container with id 4bddebc316e83fb9c8e8cc385955f8fb6add6e032f337b97ef3aef91e21944ac Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.872618 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c2a-account-create-qvx8q" event={"ID":"8933fd6b-ec1a-4c42-86a2-f09d447827b6","Type":"ContainerStarted","Data":"c19b3329a49f313e52803203d7cf38418ae83ef19a205159a1f994464af950f4"} Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.872662 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c2a-account-create-qvx8q" event={"ID":"8933fd6b-ec1a-4c42-86a2-f09d447827b6","Type":"ContainerStarted","Data":"545d8216cc4ae13b432883fa4c07fa963ee3a0d4883fe8343be214d86d7bb63b"} Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.875162 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"0ea1ffc92de0a4afc06d56088d35bbbd8a9da2d54aed518af2cab8e1b79adb8d"} Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.886864 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0102-account-create-p7v69" event={"ID":"34ce4329-745c-4427-8b49-e538401c3226","Type":"ContainerStarted","Data":"4bddebc316e83fb9c8e8cc385955f8fb6add6e032f337b97ef3aef91e21944ac"} Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.892000 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lprm4" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.891989 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lprm4" event={"ID":"c5bb6f33-61f5-4b6f-8449-2595b8593d28","Type":"ContainerDied","Data":"b1b30077cd379d3d20e30c52b561b9b793a0c3fce3568ca7a61913b50bd49b39"} Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.892457 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1b30077cd379d3d20e30c52b561b9b793a0c3fce3568ca7a61913b50bd49b39" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.893333 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-72mrw" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.893684 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-72mrw" event={"ID":"eea4fe86-b04b-44d4-bc68-361b6bf34b6a","Type":"ContainerDied","Data":"4e3263b172d403826622ad66f9a96b8978ff659e00ba02e88da794e81bf19ce6"} Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.893736 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e3263b172d403826622ad66f9a96b8978ff659e00ba02e88da794e81bf19ce6" Oct 04 04:55:08 crc kubenswrapper[4575]: I1004 04:55:08.895697 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7c2a-account-create-qvx8q" podStartSLOduration=2.895685767 podStartE2EDuration="2.895685767s" podCreationTimestamp="2025-10-04 04:55:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:08.891858417 +0000 UTC m=+1260.220417231" watchObservedRunningTime="2025-10-04 04:55:08.895685767 +0000 UTC m=+1260.224244581" Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.139405 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kjpkv" Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.172209 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxxhk\" (UniqueName: \"kubernetes.io/projected/bedd9a1e-02e4-4fac-8f24-f45234f3aa40-kube-api-access-fxxhk\") pod \"bedd9a1e-02e4-4fac-8f24-f45234f3aa40\" (UID: \"bedd9a1e-02e4-4fac-8f24-f45234f3aa40\") " Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.181390 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bedd9a1e-02e4-4fac-8f24-f45234f3aa40-kube-api-access-fxxhk" (OuterVolumeSpecName: "kube-api-access-fxxhk") pod "bedd9a1e-02e4-4fac-8f24-f45234f3aa40" (UID: "bedd9a1e-02e4-4fac-8f24-f45234f3aa40"). InnerVolumeSpecName "kube-api-access-fxxhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.274280 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxxhk\" (UniqueName: \"kubernetes.io/projected/bedd9a1e-02e4-4fac-8f24-f45234f3aa40-kube-api-access-fxxhk\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.902154 4575 generic.go:334] "Generic (PLEG): container finished" podID="8933fd6b-ec1a-4c42-86a2-f09d447827b6" containerID="c19b3329a49f313e52803203d7cf38418ae83ef19a205159a1f994464af950f4" exitCode=0 Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.902247 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c2a-account-create-qvx8q" event={"ID":"8933fd6b-ec1a-4c42-86a2-f09d447827b6","Type":"ContainerDied","Data":"c19b3329a49f313e52803203d7cf38418ae83ef19a205159a1f994464af950f4"} Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.904621 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"0ec23132eba9e85bff8cf8da26f7fed7877b36e1a966c1b139da9f062641ac5a"} Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.904653 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"a335c74738ab8f2f80afe70f428f718cb3dd0276647202b91ff6383cb81d0c50"} Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.904665 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"b139b8a7c95d8baae124e7457cc3d5bea24a93c4dd8507b89be5a6ddeacd572c"} Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.906050 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kjpkv" event={"ID":"bedd9a1e-02e4-4fac-8f24-f45234f3aa40","Type":"ContainerDied","Data":"2e8f9faa5adc35c503e37f3bf873ce1c2bcab09fb2fcb4d2e8e8638f6fdd1274"} Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.906076 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e8f9faa5adc35c503e37f3bf873ce1c2bcab09fb2fcb4d2e8e8638f6fdd1274" Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.906087 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kjpkv" Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.907870 4575 generic.go:334] "Generic (PLEG): container finished" podID="34ce4329-745c-4427-8b49-e538401c3226" containerID="d3369085759623f0837e1c0a2f233a25583887366e0576693ac6a2d4465abcc0" exitCode=0 Oct 04 04:55:09 crc kubenswrapper[4575]: I1004 04:55:09.907916 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0102-account-create-p7v69" event={"ID":"34ce4329-745c-4427-8b49-e538401c3226","Type":"ContainerDied","Data":"d3369085759623f0837e1c0a2f233a25583887366e0576693ac6a2d4465abcc0"} Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.338820 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0102-account-create-p7v69" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.346962 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c2a-account-create-qvx8q" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.425258 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbjbh\" (UniqueName: \"kubernetes.io/projected/34ce4329-745c-4427-8b49-e538401c3226-kube-api-access-lbjbh\") pod \"34ce4329-745c-4427-8b49-e538401c3226\" (UID: \"34ce4329-745c-4427-8b49-e538401c3226\") " Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.425307 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mkxk\" (UniqueName: \"kubernetes.io/projected/8933fd6b-ec1a-4c42-86a2-f09d447827b6-kube-api-access-9mkxk\") pod \"8933fd6b-ec1a-4c42-86a2-f09d447827b6\" (UID: \"8933fd6b-ec1a-4c42-86a2-f09d447827b6\") " Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.434450 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ce4329-745c-4427-8b49-e538401c3226-kube-api-access-lbjbh" (OuterVolumeSpecName: "kube-api-access-lbjbh") pod "34ce4329-745c-4427-8b49-e538401c3226" (UID: "34ce4329-745c-4427-8b49-e538401c3226"). InnerVolumeSpecName "kube-api-access-lbjbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.436798 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8933fd6b-ec1a-4c42-86a2-f09d447827b6-kube-api-access-9mkxk" (OuterVolumeSpecName: "kube-api-access-9mkxk") pod "8933fd6b-ec1a-4c42-86a2-f09d447827b6" (UID: "8933fd6b-ec1a-4c42-86a2-f09d447827b6"). InnerVolumeSpecName "kube-api-access-9mkxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.454333 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.540134 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbjbh\" (UniqueName: \"kubernetes.io/projected/34ce4329-745c-4427-8b49-e538401c3226-kube-api-access-lbjbh\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.540175 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mkxk\" (UniqueName: \"kubernetes.io/projected/8933fd6b-ec1a-4c42-86a2-f09d447827b6-kube-api-access-9mkxk\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.931866 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0102-account-create-p7v69" event={"ID":"34ce4329-745c-4427-8b49-e538401c3226","Type":"ContainerDied","Data":"4bddebc316e83fb9c8e8cc385955f8fb6add6e032f337b97ef3aef91e21944ac"} Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.932359 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bddebc316e83fb9c8e8cc385955f8fb6add6e032f337b97ef3aef91e21944ac" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.932500 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0102-account-create-p7v69" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.934378 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c2a-account-create-qvx8q" event={"ID":"8933fd6b-ec1a-4c42-86a2-f09d447827b6","Type":"ContainerDied","Data":"545d8216cc4ae13b432883fa4c07fa963ee3a0d4883fe8343be214d86d7bb63b"} Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.934423 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="545d8216cc4ae13b432883fa4c07fa963ee3a0d4883fe8343be214d86d7bb63b" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.934501 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c2a-account-create-qvx8q" Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.947116 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"27a86a0ef8c360f5234831c52be79b193f98e160d1f824bf33a8cce7600f20ee"} Oct 04 04:55:11 crc kubenswrapper[4575]: I1004 04:55:11.947158 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"d5e3894ad3c2ae893926bc3dc0270e92da85ae8ef4d5fa58c89733471e89e720"} Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.583381 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xmg65"] Oct 04 04:55:12 crc kubenswrapper[4575]: E1004 04:55:12.583796 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ce4329-745c-4427-8b49-e538401c3226" containerName="mariadb-account-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.583813 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ce4329-745c-4427-8b49-e538401c3226" containerName="mariadb-account-create" Oct 04 04:55:12 crc kubenswrapper[4575]: E1004 04:55:12.583824 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eea4fe86-b04b-44d4-bc68-361b6bf34b6a" containerName="mariadb-database-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.583831 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="eea4fe86-b04b-44d4-bc68-361b6bf34b6a" containerName="mariadb-database-create" Oct 04 04:55:12 crc kubenswrapper[4575]: E1004 04:55:12.583852 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bedd9a1e-02e4-4fac-8f24-f45234f3aa40" containerName="mariadb-database-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.583861 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="bedd9a1e-02e4-4fac-8f24-f45234f3aa40" containerName="mariadb-database-create" Oct 04 04:55:12 crc kubenswrapper[4575]: E1004 04:55:12.583873 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bb6f33-61f5-4b6f-8449-2595b8593d28" containerName="mariadb-database-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.583880 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bb6f33-61f5-4b6f-8449-2595b8593d28" containerName="mariadb-database-create" Oct 04 04:55:12 crc kubenswrapper[4575]: E1004 04:55:12.583901 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8933fd6b-ec1a-4c42-86a2-f09d447827b6" containerName="mariadb-account-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.583908 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8933fd6b-ec1a-4c42-86a2-f09d447827b6" containerName="mariadb-account-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.584077 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="8933fd6b-ec1a-4c42-86a2-f09d447827b6" containerName="mariadb-account-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.584120 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="eea4fe86-b04b-44d4-bc68-361b6bf34b6a" containerName="mariadb-database-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.584142 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bb6f33-61f5-4b6f-8449-2595b8593d28" containerName="mariadb-database-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.584156 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ce4329-745c-4427-8b49-e538401c3226" containerName="mariadb-account-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.584172 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="bedd9a1e-02e4-4fac-8f24-f45234f3aa40" containerName="mariadb-database-create" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.584811 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.593206 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.593386 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.593510 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xdgq4" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.599172 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xmg65"] Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.599877 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.659841 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-config-data\") pod \"keystone-db-sync-xmg65\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.660140 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t76ts\" (UniqueName: \"kubernetes.io/projected/d0064681-6434-46c4-b46b-2c82fc5694d0-kube-api-access-t76ts\") pod \"keystone-db-sync-xmg65\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.660296 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-combined-ca-bundle\") pod \"keystone-db-sync-xmg65\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.729194 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-66da-account-create-6ckcq"] Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.730437 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-66da-account-create-6ckcq" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.733228 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.741066 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-66da-account-create-6ckcq"] Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.763453 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-config-data\") pod \"keystone-db-sync-xmg65\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.763529 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t76ts\" (UniqueName: \"kubernetes.io/projected/d0064681-6434-46c4-b46b-2c82fc5694d0-kube-api-access-t76ts\") pod \"keystone-db-sync-xmg65\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.763606 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-combined-ca-bundle\") pod \"keystone-db-sync-xmg65\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.774506 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-combined-ca-bundle\") pod \"keystone-db-sync-xmg65\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.778474 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-config-data\") pod \"keystone-db-sync-xmg65\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.778948 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t76ts\" (UniqueName: \"kubernetes.io/projected/d0064681-6434-46c4-b46b-2c82fc5694d0-kube-api-access-t76ts\") pod \"keystone-db-sync-xmg65\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.865351 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtk2d\" (UniqueName: \"kubernetes.io/projected/bb097de0-50a2-44a0-a1b9-3abcb12741de-kube-api-access-xtk2d\") pod \"glance-66da-account-create-6ckcq\" (UID: \"bb097de0-50a2-44a0-a1b9-3abcb12741de\") " pod="openstack/glance-66da-account-create-6ckcq" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.900898 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.967416 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtk2d\" (UniqueName: \"kubernetes.io/projected/bb097de0-50a2-44a0-a1b9-3abcb12741de-kube-api-access-xtk2d\") pod \"glance-66da-account-create-6ckcq\" (UID: \"bb097de0-50a2-44a0-a1b9-3abcb12741de\") " pod="openstack/glance-66da-account-create-6ckcq" Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.981038 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"48d5e9f3c0833ea58d3c958ffe926af4e3beac4e1fe38ec36a21289c8a3ccdfa"} Oct 04 04:55:12 crc kubenswrapper[4575]: I1004 04:55:12.981387 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"39aa9655a9d75d876b36cda556aa4a38d6d38ae8a573fc76589f40ebf05aa49c"} Oct 04 04:55:13 crc kubenswrapper[4575]: I1004 04:55:13.007920 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtk2d\" (UniqueName: \"kubernetes.io/projected/bb097de0-50a2-44a0-a1b9-3abcb12741de-kube-api-access-xtk2d\") pod \"glance-66da-account-create-6ckcq\" (UID: \"bb097de0-50a2-44a0-a1b9-3abcb12741de\") " pod="openstack/glance-66da-account-create-6ckcq" Oct 04 04:55:13 crc kubenswrapper[4575]: I1004 04:55:13.049684 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-66da-account-create-6ckcq" Oct 04 04:55:13 crc kubenswrapper[4575]: I1004 04:55:13.615479 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xmg65"] Oct 04 04:55:13 crc kubenswrapper[4575]: I1004 04:55:13.693306 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-66da-account-create-6ckcq"] Oct 04 04:55:13 crc kubenswrapper[4575]: I1004 04:55:13.992266 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xmg65" event={"ID":"d0064681-6434-46c4-b46b-2c82fc5694d0","Type":"ContainerStarted","Data":"4d0f417f55f9f21d1612e0a2546ea66c415205f92e319d97ec6fff0a73a03a78"} Oct 04 04:55:13 crc kubenswrapper[4575]: I1004 04:55:13.998196 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-66da-account-create-6ckcq" event={"ID":"bb097de0-50a2-44a0-a1b9-3abcb12741de","Type":"ContainerStarted","Data":"8a62fa13e09c631a28687e22ad3927fe416005701ec969d52fd561341dfc5434"} Oct 04 04:55:13 crc kubenswrapper[4575]: I1004 04:55:13.998245 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-66da-account-create-6ckcq" event={"ID":"bb097de0-50a2-44a0-a1b9-3abcb12741de","Type":"ContainerStarted","Data":"7387d245a801a2bb6c35a27c151ece2e2686066cf69272e344300004b046520a"} Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.021378 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-66da-account-create-6ckcq" podStartSLOduration=2.021356689 podStartE2EDuration="2.021356689s" podCreationTimestamp="2025-10-04 04:55:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:14.013955556 +0000 UTC m=+1265.342514370" watchObservedRunningTime="2025-10-04 04:55:14.021356689 +0000 UTC m=+1265.349915523" Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.700751 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d9bf-account-create-hg27q"] Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.703857 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d9bf-account-create-hg27q" Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.707436 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.719369 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d9bf-account-create-hg27q"] Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.809724 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4g97\" (UniqueName: \"kubernetes.io/projected/b301bf9a-a111-4933-8f5e-51bd3d7d0c9b-kube-api-access-h4g97\") pod \"cinder-d9bf-account-create-hg27q\" (UID: \"b301bf9a-a111-4933-8f5e-51bd3d7d0c9b\") " pod="openstack/cinder-d9bf-account-create-hg27q" Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.910221 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3a4a-account-create-knhf9"] Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.911660 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4g97\" (UniqueName: \"kubernetes.io/projected/b301bf9a-a111-4933-8f5e-51bd3d7d0c9b-kube-api-access-h4g97\") pod \"cinder-d9bf-account-create-hg27q\" (UID: \"b301bf9a-a111-4933-8f5e-51bd3d7d0c9b\") " pod="openstack/cinder-d9bf-account-create-hg27q" Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.912370 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3a4a-account-create-knhf9" Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.915141 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.932338 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3a4a-account-create-knhf9"] Oct 04 04:55:14 crc kubenswrapper[4575]: I1004 04:55:14.944764 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4g97\" (UniqueName: \"kubernetes.io/projected/b301bf9a-a111-4933-8f5e-51bd3d7d0c9b-kube-api-access-h4g97\") pod \"cinder-d9bf-account-create-hg27q\" (UID: \"b301bf9a-a111-4933-8f5e-51bd3d7d0c9b\") " pod="openstack/cinder-d9bf-account-create-hg27q" Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.012087 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"75bc964c6229bb0f2b0d2cd931e8e670fae192a48b927112e94ab1391ed7d58e"} Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.012147 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"47c0025b94edf04b1e1e18cbcf273dd77522212f0349c6621b7fc6d33a38c38d"} Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.013599 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbrtg\" (UniqueName: \"kubernetes.io/projected/274f307d-9c54-49f6-9460-2231583ff053-kube-api-access-kbrtg\") pod \"barbican-3a4a-account-create-knhf9\" (UID: \"274f307d-9c54-49f6-9460-2231583ff053\") " pod="openstack/barbican-3a4a-account-create-knhf9" Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.014599 4575 generic.go:334] "Generic (PLEG): container finished" podID="bb097de0-50a2-44a0-a1b9-3abcb12741de" containerID="8a62fa13e09c631a28687e22ad3927fe416005701ec969d52fd561341dfc5434" exitCode=0 Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.014630 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-66da-account-create-6ckcq" event={"ID":"bb097de0-50a2-44a0-a1b9-3abcb12741de","Type":"ContainerDied","Data":"8a62fa13e09c631a28687e22ad3927fe416005701ec969d52fd561341dfc5434"} Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.018631 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d9bf-account-create-hg27q" Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.116325 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbrtg\" (UniqueName: \"kubernetes.io/projected/274f307d-9c54-49f6-9460-2231583ff053-kube-api-access-kbrtg\") pod \"barbican-3a4a-account-create-knhf9\" (UID: \"274f307d-9c54-49f6-9460-2231583ff053\") " pod="openstack/barbican-3a4a-account-create-knhf9" Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.144992 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbrtg\" (UniqueName: \"kubernetes.io/projected/274f307d-9c54-49f6-9460-2231583ff053-kube-api-access-kbrtg\") pod \"barbican-3a4a-account-create-knhf9\" (UID: \"274f307d-9c54-49f6-9460-2231583ff053\") " pod="openstack/barbican-3a4a-account-create-knhf9" Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.244209 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3a4a-account-create-knhf9" Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.546245 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d9bf-account-create-hg27q"] Oct 04 04:55:15 crc kubenswrapper[4575]: W1004 04:55:15.560347 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb301bf9a_a111_4933_8f5e_51bd3d7d0c9b.slice/crio-2345fbc0a501b8d60e21892ef3e11c13f7f7a6d695973cd41b69f252d4093a50 WatchSource:0}: Error finding container 2345fbc0a501b8d60e21892ef3e11c13f7f7a6d695973cd41b69f252d4093a50: Status 404 returned error can't find the container with id 2345fbc0a501b8d60e21892ef3e11c13f7f7a6d695973cd41b69f252d4093a50 Oct 04 04:55:15 crc kubenswrapper[4575]: I1004 04:55:15.763729 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3a4a-account-create-knhf9"] Oct 04 04:55:15 crc kubenswrapper[4575]: W1004 04:55:15.772423 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod274f307d_9c54_49f6_9460_2231583ff053.slice/crio-614c330e5a29a181f7071d7b92e4af4fcb7a33e1f7a6b071c06e46917045b007 WatchSource:0}: Error finding container 614c330e5a29a181f7071d7b92e4af4fcb7a33e1f7a6b071c06e46917045b007: Status 404 returned error can't find the container with id 614c330e5a29a181f7071d7b92e4af4fcb7a33e1f7a6b071c06e46917045b007 Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.030698 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d9bf-account-create-hg27q" event={"ID":"b301bf9a-a111-4933-8f5e-51bd3d7d0c9b","Type":"ContainerStarted","Data":"4c258d1f1e011ccd5f32f3add9c872c5fc1b75422008487ab35ce83605ca4137"} Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.030742 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d9bf-account-create-hg27q" event={"ID":"b301bf9a-a111-4933-8f5e-51bd3d7d0c9b","Type":"ContainerStarted","Data":"2345fbc0a501b8d60e21892ef3e11c13f7f7a6d695973cd41b69f252d4093a50"} Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.037741 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3a4a-account-create-knhf9" event={"ID":"274f307d-9c54-49f6-9460-2231583ff053","Type":"ContainerStarted","Data":"630e8523897b9da80ee6851314cb5db2e25064a70d6c3f27acf4a8aa7dc057bb"} Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.037973 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3a4a-account-create-knhf9" event={"ID":"274f307d-9c54-49f6-9460-2231583ff053","Type":"ContainerStarted","Data":"614c330e5a29a181f7071d7b92e4af4fcb7a33e1f7a6b071c06e46917045b007"} Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.050518 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"de208a0027ced4b06dc0dd1ac1e4aab478ace263b59dfe0f7891740843917fae"} Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.050575 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"b3d9e63ea0b6a87244d2481246410aa381540465239d2e38bc09b352b9b1ea9f"} Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.067222 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-d9bf-account-create-hg27q" podStartSLOduration=2.067202076 podStartE2EDuration="2.067202076s" podCreationTimestamp="2025-10-04 04:55:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:16.048424966 +0000 UTC m=+1267.376983790" watchObservedRunningTime="2025-10-04 04:55:16.067202076 +0000 UTC m=+1267.395760890" Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.071828 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-3a4a-account-create-knhf9" podStartSLOduration=2.071817319 podStartE2EDuration="2.071817319s" podCreationTimestamp="2025-10-04 04:55:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:16.065821767 +0000 UTC m=+1267.394380601" watchObservedRunningTime="2025-10-04 04:55:16.071817319 +0000 UTC m=+1267.400376133" Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.343024 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-66da-account-create-6ckcq" Oct 04 04:55:16 crc kubenswrapper[4575]: E1004 04:55:16.392725 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod274f307d_9c54_49f6_9460_2231583ff053.slice/crio-630e8523897b9da80ee6851314cb5db2e25064a70d6c3f27acf4a8aa7dc057bb.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.441836 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtk2d\" (UniqueName: \"kubernetes.io/projected/bb097de0-50a2-44a0-a1b9-3abcb12741de-kube-api-access-xtk2d\") pod \"bb097de0-50a2-44a0-a1b9-3abcb12741de\" (UID: \"bb097de0-50a2-44a0-a1b9-3abcb12741de\") " Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.460095 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb097de0-50a2-44a0-a1b9-3abcb12741de-kube-api-access-xtk2d" (OuterVolumeSpecName: "kube-api-access-xtk2d") pod "bb097de0-50a2-44a0-a1b9-3abcb12741de" (UID: "bb097de0-50a2-44a0-a1b9-3abcb12741de"). InnerVolumeSpecName "kube-api-access-xtk2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:16 crc kubenswrapper[4575]: I1004 04:55:16.543528 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtk2d\" (UniqueName: \"kubernetes.io/projected/bb097de0-50a2-44a0-a1b9-3abcb12741de-kube-api-access-xtk2d\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.065835 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"c0b9e92c00298a1b1cb4d2d30efbd9ee2d2aa0b8f80845091189ad785e4de18c"} Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.066161 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"321558c5e29cd5ae72f654e12ee302292b6f03fa66a7897cc855c1de655ecb58"} Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.066174 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c9ec5625-8b15-42bd-8258-4960d8000469","Type":"ContainerStarted","Data":"e1ace6611dec114fe28fa98ae1ef962a8c7309a8a941e5f867be9d524cdad63f"} Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.071805 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-66da-account-create-6ckcq" event={"ID":"bb097de0-50a2-44a0-a1b9-3abcb12741de","Type":"ContainerDied","Data":"7387d245a801a2bb6c35a27c151ece2e2686066cf69272e344300004b046520a"} Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.071848 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7387d245a801a2bb6c35a27c151ece2e2686066cf69272e344300004b046520a" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.071850 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-66da-account-create-6ckcq" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.073237 4575 generic.go:334] "Generic (PLEG): container finished" podID="b301bf9a-a111-4933-8f5e-51bd3d7d0c9b" containerID="4c258d1f1e011ccd5f32f3add9c872c5fc1b75422008487ab35ce83605ca4137" exitCode=0 Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.073299 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d9bf-account-create-hg27q" event={"ID":"b301bf9a-a111-4933-8f5e-51bd3d7d0c9b","Type":"ContainerDied","Data":"4c258d1f1e011ccd5f32f3add9c872c5fc1b75422008487ab35ce83605ca4137"} Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.076899 4575 generic.go:334] "Generic (PLEG): container finished" podID="274f307d-9c54-49f6-9460-2231583ff053" containerID="630e8523897b9da80ee6851314cb5db2e25064a70d6c3f27acf4a8aa7dc057bb" exitCode=0 Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.076956 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3a4a-account-create-knhf9" event={"ID":"274f307d-9c54-49f6-9460-2231583ff053","Type":"ContainerDied","Data":"630e8523897b9da80ee6851314cb5db2e25064a70d6c3f27acf4a8aa7dc057bb"} Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.109332 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.856285127 podStartE2EDuration="47.109311218s" podCreationTimestamp="2025-10-04 04:54:30 +0000 UTC" firstStartedPulling="2025-10-04 04:55:04.118944082 +0000 UTC m=+1255.447502886" lastFinishedPulling="2025-10-04 04:55:14.371970163 +0000 UTC m=+1265.700528977" observedRunningTime="2025-10-04 04:55:17.102223574 +0000 UTC m=+1268.430782418" watchObservedRunningTime="2025-10-04 04:55:17.109311218 +0000 UTC m=+1268.437870022" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.418312 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-ttlxx"] Oct 04 04:55:17 crc kubenswrapper[4575]: E1004 04:55:17.418736 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb097de0-50a2-44a0-a1b9-3abcb12741de" containerName="mariadb-account-create" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.418753 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb097de0-50a2-44a0-a1b9-3abcb12741de" containerName="mariadb-account-create" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.418950 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb097de0-50a2-44a0-a1b9-3abcb12741de" containerName="mariadb-account-create" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.419855 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.421748 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.460260 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.460328 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsdnb\" (UniqueName: \"kubernetes.io/projected/6dca86a7-57a1-457e-adf9-58882540f233-kube-api-access-vsdnb\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.460353 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-config\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.460435 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.460474 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.460503 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.479899 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-ttlxx"] Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.561785 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.561843 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsdnb\" (UniqueName: \"kubernetes.io/projected/6dca86a7-57a1-457e-adf9-58882540f233-kube-api-access-vsdnb\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.561873 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-config\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.561926 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.561957 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.561976 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.563026 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.563575 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.563738 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-config\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.564041 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.564103 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.594060 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsdnb\" (UniqueName: \"kubernetes.io/projected/6dca86a7-57a1-457e-adf9-58882540f233-kube-api-access-vsdnb\") pod \"dnsmasq-dns-6d5b6d6b67-ttlxx\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.738826 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.869848 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-gtvfg"] Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.871797 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.878325 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xtpfm" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.878373 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.893938 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-gtvfg"] Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.973796 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-combined-ca-bundle\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.973904 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-db-sync-config-data\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.973966 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hltsv\" (UniqueName: \"kubernetes.io/projected/775d96cb-5e3b-437f-b61d-7799e86f2f58-kube-api-access-hltsv\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:17 crc kubenswrapper[4575]: I1004 04:55:17.974089 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-config-data\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:18 crc kubenswrapper[4575]: I1004 04:55:18.076179 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-db-sync-config-data\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:18 crc kubenswrapper[4575]: I1004 04:55:18.076265 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hltsv\" (UniqueName: \"kubernetes.io/projected/775d96cb-5e3b-437f-b61d-7799e86f2f58-kube-api-access-hltsv\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:18 crc kubenswrapper[4575]: I1004 04:55:18.076378 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-config-data\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:18 crc kubenswrapper[4575]: I1004 04:55:18.076448 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-combined-ca-bundle\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:18 crc kubenswrapper[4575]: I1004 04:55:18.080150 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-config-data\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:18 crc kubenswrapper[4575]: I1004 04:55:18.098769 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-db-sync-config-data\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:18 crc kubenswrapper[4575]: I1004 04:55:18.104414 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hltsv\" (UniqueName: \"kubernetes.io/projected/775d96cb-5e3b-437f-b61d-7799e86f2f58-kube-api-access-hltsv\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:18 crc kubenswrapper[4575]: I1004 04:55:18.112990 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-combined-ca-bundle\") pod \"glance-db-sync-gtvfg\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:18 crc kubenswrapper[4575]: I1004 04:55:18.206495 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gtvfg" Oct 04 04:55:20 crc kubenswrapper[4575]: I1004 04:55:20.728366 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3a4a-account-create-knhf9" Oct 04 04:55:20 crc kubenswrapper[4575]: I1004 04:55:20.738429 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d9bf-account-create-hg27q" Oct 04 04:55:20 crc kubenswrapper[4575]: I1004 04:55:20.821955 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4g97\" (UniqueName: \"kubernetes.io/projected/b301bf9a-a111-4933-8f5e-51bd3d7d0c9b-kube-api-access-h4g97\") pod \"b301bf9a-a111-4933-8f5e-51bd3d7d0c9b\" (UID: \"b301bf9a-a111-4933-8f5e-51bd3d7d0c9b\") " Oct 04 04:55:20 crc kubenswrapper[4575]: I1004 04:55:20.822091 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbrtg\" (UniqueName: \"kubernetes.io/projected/274f307d-9c54-49f6-9460-2231583ff053-kube-api-access-kbrtg\") pod \"274f307d-9c54-49f6-9460-2231583ff053\" (UID: \"274f307d-9c54-49f6-9460-2231583ff053\") " Oct 04 04:55:20 crc kubenswrapper[4575]: I1004 04:55:20.830942 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b301bf9a-a111-4933-8f5e-51bd3d7d0c9b-kube-api-access-h4g97" (OuterVolumeSpecName: "kube-api-access-h4g97") pod "b301bf9a-a111-4933-8f5e-51bd3d7d0c9b" (UID: "b301bf9a-a111-4933-8f5e-51bd3d7d0c9b"). InnerVolumeSpecName "kube-api-access-h4g97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:20 crc kubenswrapper[4575]: I1004 04:55:20.843965 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/274f307d-9c54-49f6-9460-2231583ff053-kube-api-access-kbrtg" (OuterVolumeSpecName: "kube-api-access-kbrtg") pod "274f307d-9c54-49f6-9460-2231583ff053" (UID: "274f307d-9c54-49f6-9460-2231583ff053"). InnerVolumeSpecName "kube-api-access-kbrtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:20 crc kubenswrapper[4575]: I1004 04:55:20.926575 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbrtg\" (UniqueName: \"kubernetes.io/projected/274f307d-9c54-49f6-9460-2231583ff053-kube-api-access-kbrtg\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:20 crc kubenswrapper[4575]: I1004 04:55:20.926627 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4g97\" (UniqueName: \"kubernetes.io/projected/b301bf9a-a111-4933-8f5e-51bd3d7d0c9b-kube-api-access-h4g97\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:21 crc kubenswrapper[4575]: I1004 04:55:21.079785 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-ttlxx"] Oct 04 04:55:21 crc kubenswrapper[4575]: W1004 04:55:21.081255 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dca86a7_57a1_457e_adf9_58882540f233.slice/crio-ef2c4f1f74dd6cd838724c79409e3d78d6f01ecf7ff654427bea017764bf3041 WatchSource:0}: Error finding container ef2c4f1f74dd6cd838724c79409e3d78d6f01ecf7ff654427bea017764bf3041: Status 404 returned error can't find the container with id ef2c4f1f74dd6cd838724c79409e3d78d6f01ecf7ff654427bea017764bf3041 Oct 04 04:55:21 crc kubenswrapper[4575]: I1004 04:55:21.120222 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d9bf-account-create-hg27q" event={"ID":"b301bf9a-a111-4933-8f5e-51bd3d7d0c9b","Type":"ContainerDied","Data":"2345fbc0a501b8d60e21892ef3e11c13f7f7a6d695973cd41b69f252d4093a50"} Oct 04 04:55:21 crc kubenswrapper[4575]: I1004 04:55:21.120267 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2345fbc0a501b8d60e21892ef3e11c13f7f7a6d695973cd41b69f252d4093a50" Oct 04 04:55:21 crc kubenswrapper[4575]: I1004 04:55:21.120354 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d9bf-account-create-hg27q" Oct 04 04:55:21 crc kubenswrapper[4575]: I1004 04:55:21.123879 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3a4a-account-create-knhf9" event={"ID":"274f307d-9c54-49f6-9460-2231583ff053","Type":"ContainerDied","Data":"614c330e5a29a181f7071d7b92e4af4fcb7a33e1f7a6b071c06e46917045b007"} Oct 04 04:55:21 crc kubenswrapper[4575]: I1004 04:55:21.123929 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="614c330e5a29a181f7071d7b92e4af4fcb7a33e1f7a6b071c06e46917045b007" Oct 04 04:55:21 crc kubenswrapper[4575]: I1004 04:55:21.124011 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3a4a-account-create-knhf9" Oct 04 04:55:21 crc kubenswrapper[4575]: I1004 04:55:21.127814 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" event={"ID":"6dca86a7-57a1-457e-adf9-58882540f233","Type":"ContainerStarted","Data":"ef2c4f1f74dd6cd838724c79409e3d78d6f01ecf7ff654427bea017764bf3041"} Oct 04 04:55:21 crc kubenswrapper[4575]: I1004 04:55:21.210867 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-gtvfg"] Oct 04 04:55:22 crc kubenswrapper[4575]: I1004 04:55:22.136203 4575 generic.go:334] "Generic (PLEG): container finished" podID="6dca86a7-57a1-457e-adf9-58882540f233" containerID="f1617c5123c746de83fa90f0e4e7e10b3e84efd6bd916380ea9cd2369ae67832" exitCode=0 Oct 04 04:55:22 crc kubenswrapper[4575]: I1004 04:55:22.136674 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" event={"ID":"6dca86a7-57a1-457e-adf9-58882540f233","Type":"ContainerDied","Data":"f1617c5123c746de83fa90f0e4e7e10b3e84efd6bd916380ea9cd2369ae67832"} Oct 04 04:55:22 crc kubenswrapper[4575]: I1004 04:55:22.142025 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gtvfg" event={"ID":"775d96cb-5e3b-437f-b61d-7799e86f2f58","Type":"ContainerStarted","Data":"dfcbe4cddb62c444eef9edf121c5a3903a63dd2d748a3393dd625afbcbee76fb"} Oct 04 04:55:22 crc kubenswrapper[4575]: I1004 04:55:22.146748 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xmg65" event={"ID":"d0064681-6434-46c4-b46b-2c82fc5694d0","Type":"ContainerStarted","Data":"8e745803e96988e1779f47cccad8a0e0cc731d7fe71b13770fabc2425ec358c1"} Oct 04 04:55:22 crc kubenswrapper[4575]: I1004 04:55:22.189720 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xmg65" podStartSLOduration=2.852845747 podStartE2EDuration="10.189701646s" podCreationTimestamp="2025-10-04 04:55:12 +0000 UTC" firstStartedPulling="2025-10-04 04:55:13.619547302 +0000 UTC m=+1264.948106116" lastFinishedPulling="2025-10-04 04:55:20.956403201 +0000 UTC m=+1272.284962015" observedRunningTime="2025-10-04 04:55:22.184448495 +0000 UTC m=+1273.513007309" watchObservedRunningTime="2025-10-04 04:55:22.189701646 +0000 UTC m=+1273.518260460" Oct 04 04:55:23 crc kubenswrapper[4575]: I1004 04:55:23.160092 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" event={"ID":"6dca86a7-57a1-457e-adf9-58882540f233","Type":"ContainerStarted","Data":"da60a53726dd43ff90d13c3b3f93974566e2f3b78dfb1904ec524d845f90485e"} Oct 04 04:55:23 crc kubenswrapper[4575]: I1004 04:55:23.192787 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" podStartSLOduration=6.192769812 podStartE2EDuration="6.192769812s" podCreationTimestamp="2025-10-04 04:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:23.186637786 +0000 UTC m=+1274.515196610" watchObservedRunningTime="2025-10-04 04:55:23.192769812 +0000 UTC m=+1274.521328626" Oct 04 04:55:24 crc kubenswrapper[4575]: I1004 04:55:24.169027 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.139359 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-191e-account-create-62c26"] Oct 04 04:55:25 crc kubenswrapper[4575]: E1004 04:55:25.140046 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b301bf9a-a111-4933-8f5e-51bd3d7d0c9b" containerName="mariadb-account-create" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.140066 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="b301bf9a-a111-4933-8f5e-51bd3d7d0c9b" containerName="mariadb-account-create" Oct 04 04:55:25 crc kubenswrapper[4575]: E1004 04:55:25.140094 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274f307d-9c54-49f6-9460-2231583ff053" containerName="mariadb-account-create" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.140101 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="274f307d-9c54-49f6-9460-2231583ff053" containerName="mariadb-account-create" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.140239 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="b301bf9a-a111-4933-8f5e-51bd3d7d0c9b" containerName="mariadb-account-create" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.140259 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="274f307d-9c54-49f6-9460-2231583ff053" containerName="mariadb-account-create" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.142484 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-191e-account-create-62c26" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.145703 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.148301 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-191e-account-create-62c26"] Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.206753 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbr77\" (UniqueName: \"kubernetes.io/projected/918e18c5-12f4-4bf4-97c1-8bcb8441664b-kube-api-access-bbr77\") pod \"neutron-191e-account-create-62c26\" (UID: \"918e18c5-12f4-4bf4-97c1-8bcb8441664b\") " pod="openstack/neutron-191e-account-create-62c26" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.309257 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbr77\" (UniqueName: \"kubernetes.io/projected/918e18c5-12f4-4bf4-97c1-8bcb8441664b-kube-api-access-bbr77\") pod \"neutron-191e-account-create-62c26\" (UID: \"918e18c5-12f4-4bf4-97c1-8bcb8441664b\") " pod="openstack/neutron-191e-account-create-62c26" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.333696 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbr77\" (UniqueName: \"kubernetes.io/projected/918e18c5-12f4-4bf4-97c1-8bcb8441664b-kube-api-access-bbr77\") pod \"neutron-191e-account-create-62c26\" (UID: \"918e18c5-12f4-4bf4-97c1-8bcb8441664b\") " pod="openstack/neutron-191e-account-create-62c26" Oct 04 04:55:25 crc kubenswrapper[4575]: I1004 04:55:25.461166 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-191e-account-create-62c26" Oct 04 04:55:26 crc kubenswrapper[4575]: I1004 04:55:25.945913 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-191e-account-create-62c26"] Oct 04 04:55:26 crc kubenswrapper[4575]: W1004 04:55:25.953778 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod918e18c5_12f4_4bf4_97c1_8bcb8441664b.slice/crio-f4194020cf75050e3bd685bd370e7f855acc78dcbd7a8684f0ab079fbf71bf78 WatchSource:0}: Error finding container f4194020cf75050e3bd685bd370e7f855acc78dcbd7a8684f0ab079fbf71bf78: Status 404 returned error can't find the container with id f4194020cf75050e3bd685bd370e7f855acc78dcbd7a8684f0ab079fbf71bf78 Oct 04 04:55:26 crc kubenswrapper[4575]: I1004 04:55:26.195759 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-191e-account-create-62c26" event={"ID":"918e18c5-12f4-4bf4-97c1-8bcb8441664b","Type":"ContainerStarted","Data":"e716ad0d9bb072bfce3ab88e92d73af8a20f5e11423c4c218f2b0d79a1c43ca7"} Oct 04 04:55:26 crc kubenswrapper[4575]: I1004 04:55:26.195830 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-191e-account-create-62c26" event={"ID":"918e18c5-12f4-4bf4-97c1-8bcb8441664b","Type":"ContainerStarted","Data":"f4194020cf75050e3bd685bd370e7f855acc78dcbd7a8684f0ab079fbf71bf78"} Oct 04 04:55:26 crc kubenswrapper[4575]: I1004 04:55:26.223371 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-191e-account-create-62c26" podStartSLOduration=1.22334546 podStartE2EDuration="1.22334546s" podCreationTimestamp="2025-10-04 04:55:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:26.217294275 +0000 UTC m=+1277.545853089" watchObservedRunningTime="2025-10-04 04:55:26.22334546 +0000 UTC m=+1277.551904284" Oct 04 04:55:27 crc kubenswrapper[4575]: I1004 04:55:27.206890 4575 generic.go:334] "Generic (PLEG): container finished" podID="918e18c5-12f4-4bf4-97c1-8bcb8441664b" containerID="e716ad0d9bb072bfce3ab88e92d73af8a20f5e11423c4c218f2b0d79a1c43ca7" exitCode=0 Oct 04 04:55:27 crc kubenswrapper[4575]: I1004 04:55:27.206975 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-191e-account-create-62c26" event={"ID":"918e18c5-12f4-4bf4-97c1-8bcb8441664b","Type":"ContainerDied","Data":"e716ad0d9bb072bfce3ab88e92d73af8a20f5e11423c4c218f2b0d79a1c43ca7"} Oct 04 04:55:27 crc kubenswrapper[4575]: I1004 04:55:27.741739 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:55:27 crc kubenswrapper[4575]: I1004 04:55:27.807539 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gm5df"] Oct 04 04:55:27 crc kubenswrapper[4575]: I1004 04:55:27.807835 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" podUID="12365747-76ba-4bcb-a828-147712279d49" containerName="dnsmasq-dns" containerID="cri-o://67dc4c51cb6b0f7f70e31d30139b282d6140c7bc3e753f41ab9db4cfd414380e" gracePeriod=10 Oct 04 04:55:28 crc kubenswrapper[4575]: I1004 04:55:28.226631 4575 generic.go:334] "Generic (PLEG): container finished" podID="d0064681-6434-46c4-b46b-2c82fc5694d0" containerID="8e745803e96988e1779f47cccad8a0e0cc731d7fe71b13770fabc2425ec358c1" exitCode=0 Oct 04 04:55:28 crc kubenswrapper[4575]: I1004 04:55:28.226740 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xmg65" event={"ID":"d0064681-6434-46c4-b46b-2c82fc5694d0","Type":"ContainerDied","Data":"8e745803e96988e1779f47cccad8a0e0cc731d7fe71b13770fabc2425ec358c1"} Oct 04 04:55:28 crc kubenswrapper[4575]: I1004 04:55:28.230212 4575 generic.go:334] "Generic (PLEG): container finished" podID="12365747-76ba-4bcb-a828-147712279d49" containerID="67dc4c51cb6b0f7f70e31d30139b282d6140c7bc3e753f41ab9db4cfd414380e" exitCode=0 Oct 04 04:55:28 crc kubenswrapper[4575]: I1004 04:55:28.230470 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" event={"ID":"12365747-76ba-4bcb-a828-147712279d49","Type":"ContainerDied","Data":"67dc4c51cb6b0f7f70e31d30139b282d6140c7bc3e753f41ab9db4cfd414380e"} Oct 04 04:55:29 crc kubenswrapper[4575]: I1004 04:55:29.090067 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" podUID="12365747-76ba-4bcb-a828-147712279d49" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.090408 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" podUID="12365747-76ba-4bcb-a828-147712279d49" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.268250 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.273916 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-191e-account-create-62c26" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.308444 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xmg65" event={"ID":"d0064681-6434-46c4-b46b-2c82fc5694d0","Type":"ContainerDied","Data":"4d0f417f55f9f21d1612e0a2546ea66c415205f92e319d97ec6fff0a73a03a78"} Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.308510 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d0f417f55f9f21d1612e0a2546ea66c415205f92e319d97ec6fff0a73a03a78" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.308694 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xmg65" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.339388 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-191e-account-create-62c26" event={"ID":"918e18c5-12f4-4bf4-97c1-8bcb8441664b","Type":"ContainerDied","Data":"f4194020cf75050e3bd685bd370e7f855acc78dcbd7a8684f0ab079fbf71bf78"} Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.339516 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4194020cf75050e3bd685bd370e7f855acc78dcbd7a8684f0ab079fbf71bf78" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.339464 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-191e-account-create-62c26" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.373888 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-combined-ca-bundle\") pod \"d0064681-6434-46c4-b46b-2c82fc5694d0\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.374932 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbr77\" (UniqueName: \"kubernetes.io/projected/918e18c5-12f4-4bf4-97c1-8bcb8441664b-kube-api-access-bbr77\") pod \"918e18c5-12f4-4bf4-97c1-8bcb8441664b\" (UID: \"918e18c5-12f4-4bf4-97c1-8bcb8441664b\") " Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.375040 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t76ts\" (UniqueName: \"kubernetes.io/projected/d0064681-6434-46c4-b46b-2c82fc5694d0-kube-api-access-t76ts\") pod \"d0064681-6434-46c4-b46b-2c82fc5694d0\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.375117 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-config-data\") pod \"d0064681-6434-46c4-b46b-2c82fc5694d0\" (UID: \"d0064681-6434-46c4-b46b-2c82fc5694d0\") " Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.383874 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0064681-6434-46c4-b46b-2c82fc5694d0-kube-api-access-t76ts" (OuterVolumeSpecName: "kube-api-access-t76ts") pod "d0064681-6434-46c4-b46b-2c82fc5694d0" (UID: "d0064681-6434-46c4-b46b-2c82fc5694d0"). InnerVolumeSpecName "kube-api-access-t76ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.397925 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/918e18c5-12f4-4bf4-97c1-8bcb8441664b-kube-api-access-bbr77" (OuterVolumeSpecName: "kube-api-access-bbr77") pod "918e18c5-12f4-4bf4-97c1-8bcb8441664b" (UID: "918e18c5-12f4-4bf4-97c1-8bcb8441664b"). InnerVolumeSpecName "kube-api-access-bbr77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.453036 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-config-data" (OuterVolumeSpecName: "config-data") pod "d0064681-6434-46c4-b46b-2c82fc5694d0" (UID: "d0064681-6434-46c4-b46b-2c82fc5694d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.472920 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0064681-6434-46c4-b46b-2c82fc5694d0" (UID: "d0064681-6434-46c4-b46b-2c82fc5694d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.476985 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.477021 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbr77\" (UniqueName: \"kubernetes.io/projected/918e18c5-12f4-4bf4-97c1-8bcb8441664b-kube-api-access-bbr77\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.477035 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t76ts\" (UniqueName: \"kubernetes.io/projected/d0064681-6434-46c4-b46b-2c82fc5694d0-kube-api-access-t76ts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.477045 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0064681-6434-46c4-b46b-2c82fc5694d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.519143 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.679920 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-dns-svc\") pod \"12365747-76ba-4bcb-a828-147712279d49\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.680029 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xll4s\" (UniqueName: \"kubernetes.io/projected/12365747-76ba-4bcb-a828-147712279d49-kube-api-access-xll4s\") pod \"12365747-76ba-4bcb-a828-147712279d49\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.680087 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-config\") pod \"12365747-76ba-4bcb-a828-147712279d49\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.680182 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-sb\") pod \"12365747-76ba-4bcb-a828-147712279d49\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.680238 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-nb\") pod \"12365747-76ba-4bcb-a828-147712279d49\" (UID: \"12365747-76ba-4bcb-a828-147712279d49\") " Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.683801 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12365747-76ba-4bcb-a828-147712279d49-kube-api-access-xll4s" (OuterVolumeSpecName: "kube-api-access-xll4s") pod "12365747-76ba-4bcb-a828-147712279d49" (UID: "12365747-76ba-4bcb-a828-147712279d49"). InnerVolumeSpecName "kube-api-access-xll4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.726951 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "12365747-76ba-4bcb-a828-147712279d49" (UID: "12365747-76ba-4bcb-a828-147712279d49"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.728387 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "12365747-76ba-4bcb-a828-147712279d49" (UID: "12365747-76ba-4bcb-a828-147712279d49"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.729999 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "12365747-76ba-4bcb-a828-147712279d49" (UID: "12365747-76ba-4bcb-a828-147712279d49"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.731835 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-config" (OuterVolumeSpecName: "config") pod "12365747-76ba-4bcb-a828-147712279d49" (UID: "12365747-76ba-4bcb-a828-147712279d49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.781725 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.781761 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.781770 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.781779 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xll4s\" (UniqueName: \"kubernetes.io/projected/12365747-76ba-4bcb-a828-147712279d49-kube-api-access-xll4s\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:34 crc kubenswrapper[4575]: I1004 04:55:34.781789 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12365747-76ba-4bcb-a828-147712279d49-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.350522 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gtvfg" event={"ID":"775d96cb-5e3b-437f-b61d-7799e86f2f58","Type":"ContainerStarted","Data":"b88c3e2d3ac777ec98e5b700d75c1cfaf4895dd17819590c14bf1ffc0030ec62"} Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.353185 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" event={"ID":"12365747-76ba-4bcb-a828-147712279d49","Type":"ContainerDied","Data":"56e379fc5f2b32ddca9caf1e1330ee1fe92f1360b64525e2318ad2306b9c6d6a"} Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.353234 4575 scope.go:117] "RemoveContainer" containerID="67dc4c51cb6b0f7f70e31d30139b282d6140c7bc3e753f41ab9db4cfd414380e" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.353352 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gm5df" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.376505 4575 scope.go:117] "RemoveContainer" containerID="fcf6b1710d40d50b14f172a04117c66e789adca0a5d14568ebc089c43098dc71" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.387856 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-gtvfg" podStartSLOduration=5.127451124 podStartE2EDuration="18.387840214s" podCreationTimestamp="2025-10-04 04:55:17 +0000 UTC" firstStartedPulling="2025-10-04 04:55:21.233382655 +0000 UTC m=+1272.561941469" lastFinishedPulling="2025-10-04 04:55:34.493771745 +0000 UTC m=+1285.822330559" observedRunningTime="2025-10-04 04:55:35.372274506 +0000 UTC m=+1286.700833330" watchObservedRunningTime="2025-10-04 04:55:35.387840214 +0000 UTC m=+1286.716399038" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.392753 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gm5df"] Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.401653 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gm5df"] Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.580924 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-dkfg2"] Oct 04 04:55:35 crc kubenswrapper[4575]: E1004 04:55:35.581344 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="918e18c5-12f4-4bf4-97c1-8bcb8441664b" containerName="mariadb-account-create" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.581360 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="918e18c5-12f4-4bf4-97c1-8bcb8441664b" containerName="mariadb-account-create" Oct 04 04:55:35 crc kubenswrapper[4575]: E1004 04:55:35.581369 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0064681-6434-46c4-b46b-2c82fc5694d0" containerName="keystone-db-sync" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.581375 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0064681-6434-46c4-b46b-2c82fc5694d0" containerName="keystone-db-sync" Oct 04 04:55:35 crc kubenswrapper[4575]: E1004 04:55:35.581418 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12365747-76ba-4bcb-a828-147712279d49" containerName="dnsmasq-dns" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.581426 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="12365747-76ba-4bcb-a828-147712279d49" containerName="dnsmasq-dns" Oct 04 04:55:35 crc kubenswrapper[4575]: E1004 04:55:35.581434 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12365747-76ba-4bcb-a828-147712279d49" containerName="init" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.581439 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="12365747-76ba-4bcb-a828-147712279d49" containerName="init" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.581647 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="12365747-76ba-4bcb-a828-147712279d49" containerName="dnsmasq-dns" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.581689 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="918e18c5-12f4-4bf4-97c1-8bcb8441664b" containerName="mariadb-account-create" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.581701 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0064681-6434-46c4-b46b-2c82fc5694d0" containerName="keystone-db-sync" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.582912 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.617718 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nqbf7"] Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.619467 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.623208 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.623433 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xdgq4" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.623471 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.624028 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-dkfg2"] Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.634887 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.639810 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nqbf7"] Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.704442 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-config\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.704515 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbt7r\" (UniqueName: \"kubernetes.io/projected/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-kube-api-access-xbt7r\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.704552 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.704661 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.704686 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.704715 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.806898 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-fernet-keys\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809111 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-config\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809187 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-config-data\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809214 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbt7r\" (UniqueName: \"kubernetes.io/projected/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-kube-api-access-xbt7r\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809362 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809409 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxp57\" (UniqueName: \"kubernetes.io/projected/0584c5bf-98ef-462d-b62d-8211acd79e8e-kube-api-access-zxp57\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809454 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-scripts\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809506 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809532 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809706 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-combined-ca-bundle\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809728 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.809884 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-credential-keys\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.811682 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-config\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.813288 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.814026 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.816938 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.817942 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.876517 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbt7r\" (UniqueName: \"kubernetes.io/projected/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-kube-api-access-xbt7r\") pod \"dnsmasq-dns-6f8c45789f-dkfg2\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.902811 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d7cd44969-cjg5j"] Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.904012 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.904624 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.913787 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.914658 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-hz475" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.914806 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.914989 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.935654 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-credential-keys\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.935708 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-fernet-keys\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.935774 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-config-data\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.935815 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxp57\" (UniqueName: \"kubernetes.io/projected/0584c5bf-98ef-462d-b62d-8211acd79e8e-kube-api-access-zxp57\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.935843 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-scripts\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.935896 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-combined-ca-bundle\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.943017 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-fernet-keys\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.946117 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-config-data\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.947130 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d7cd44969-cjg5j"] Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.956833 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-scripts\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.957234 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-credential-keys\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:35 crc kubenswrapper[4575]: I1004 04:55:35.957327 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-combined-ca-bundle\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.019646 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxp57\" (UniqueName: \"kubernetes.io/projected/0584c5bf-98ef-462d-b62d-8211acd79e8e-kube-api-access-zxp57\") pod \"keystone-bootstrap-nqbf7\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.037874 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-config-data\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.038043 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwtlp\" (UniqueName: \"kubernetes.io/projected/27553b01-0eaa-4cad-96f2-afca79cbefc5-kube-api-access-zwtlp\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.038112 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27553b01-0eaa-4cad-96f2-afca79cbefc5-logs\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.038175 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-scripts\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.038205 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27553b01-0eaa-4cad-96f2-afca79cbefc5-horizon-secret-key\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.120468 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-kg5nx"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.121750 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.130362 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.130687 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.134559 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-c79lb" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.141635 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwtlp\" (UniqueName: \"kubernetes.io/projected/27553b01-0eaa-4cad-96f2-afca79cbefc5-kube-api-access-zwtlp\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.141731 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27553b01-0eaa-4cad-96f2-afca79cbefc5-logs\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.141777 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-scripts\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.141806 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27553b01-0eaa-4cad-96f2-afca79cbefc5-horizon-secret-key\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.141836 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-config-data\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.143305 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-config-data\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.144163 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-scripts\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.144432 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27553b01-0eaa-4cad-96f2-afca79cbefc5-logs\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.150069 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27553b01-0eaa-4cad-96f2-afca79cbefc5-horizon-secret-key\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.161712 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kg5nx"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.183725 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-6qj89"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.194970 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.225505 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-pmrbn" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.225834 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.226039 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6qj89"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.243078 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-scripts\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.243152 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-db-sync-config-data\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.243222 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-combined-ca-bundle\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.243240 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-config-data\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.243298 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef74f3cc-9958-4de3-a914-12e5453e169d-etc-machine-id\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.243324 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2tjm\" (UniqueName: \"kubernetes.io/projected/ef74f3cc-9958-4de3-a914-12e5453e169d-kube-api-access-s2tjm\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.244283 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwtlp\" (UniqueName: \"kubernetes.io/projected/27553b01-0eaa-4cad-96f2-afca79cbefc5-kube-api-access-zwtlp\") pod \"horizon-7d7cd44969-cjg5j\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.254404 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.257456 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d5695b9d5-jnxdq"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.263790 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.276182 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d5695b9d5-jnxdq"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.341410 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-459j5"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.342707 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.344626 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.345436 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-combined-ca-bundle\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.345687 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-config-data\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.345827 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-db-sync-config-data\") pod \"barbican-db-sync-6qj89\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.345871 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfdtp\" (UniqueName: \"kubernetes.io/projected/62f6c233-faeb-47e1-8103-4902eaea8d4a-kube-api-access-hfdtp\") pod \"barbican-db-sync-6qj89\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.345939 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef74f3cc-9958-4de3-a914-12e5453e169d-etc-machine-id\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.345998 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2tjm\" (UniqueName: \"kubernetes.io/projected/ef74f3cc-9958-4de3-a914-12e5453e169d-kube-api-access-s2tjm\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.346030 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-scripts\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.346097 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-db-sync-config-data\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.346171 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-combined-ca-bundle\") pod \"barbican-db-sync-6qj89\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.346398 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef74f3cc-9958-4de3-a914-12e5453e169d-etc-machine-id\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.355032 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.355299 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-7pkll" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.355057 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.364420 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-combined-ca-bundle\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.369448 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-scripts\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.380305 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-db-sync-config-data\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.382826 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-config-data\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.394079 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-dkfg2"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.436630 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.437057 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2tjm\" (UniqueName: \"kubernetes.io/projected/ef74f3cc-9958-4de3-a914-12e5453e169d-kube-api-access-s2tjm\") pod \"cinder-db-sync-kg5nx\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.439170 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.445071 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.445286 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447109 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-combined-ca-bundle\") pod \"barbican-db-sync-6qj89\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447144 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-config-data\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447173 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-scripts\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447202 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-combined-ca-bundle\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447247 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-scripts\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447285 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cndl\" (UniqueName: \"kubernetes.io/projected/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-kube-api-access-2cndl\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447312 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-db-sync-config-data\") pod \"barbican-db-sync-6qj89\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447337 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzfdm\" (UniqueName: \"kubernetes.io/projected/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-kube-api-access-hzfdm\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447367 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfdtp\" (UniqueName: \"kubernetes.io/projected/62f6c233-faeb-47e1-8103-4902eaea8d4a-kube-api-access-hfdtp\") pod \"barbican-db-sync-6qj89\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447388 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-horizon-secret-key\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.447412 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-logs\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.457799 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-459j5"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.463734 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-logs\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.463847 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-config-data\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.472652 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.479747 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-db-sync-config-data\") pod \"barbican-db-sync-6qj89\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.492880 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-cwbzl"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.494825 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.501467 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-combined-ca-bundle\") pod \"barbican-db-sync-6qj89\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.534445 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfdtp\" (UniqueName: \"kubernetes.io/projected/62f6c233-faeb-47e1-8103-4902eaea8d4a-kube-api-access-hfdtp\") pod \"barbican-db-sync-6qj89\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.542713 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565647 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565699 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-config-data\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565817 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-config-data\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565836 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-scripts\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565860 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-combined-ca-bundle\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565892 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-scripts\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565922 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-scripts\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565937 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565967 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cndl\" (UniqueName: \"kubernetes.io/projected/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-kube-api-access-2cndl\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.565988 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-log-httpd\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.566003 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-run-httpd\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.566022 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzfdm\" (UniqueName: \"kubernetes.io/projected/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-kube-api-access-hzfdm\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.566043 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-horizon-secret-key\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.566061 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-logs\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.566080 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpdhr\" (UniqueName: \"kubernetes.io/projected/89017745-df9d-4bae-ad06-786a6b653ce8-kube-api-access-wpdhr\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.566118 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-logs\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.566147 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-config-data\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.571032 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-config-data\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.575324 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-cwbzl"] Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.576850 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-scripts\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.577817 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-logs\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.578198 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-logs\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.587861 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-horizon-secret-key\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.610441 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-scripts\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.613105 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-config-data\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.619102 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-combined-ca-bundle\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.628320 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzfdm\" (UniqueName: \"kubernetes.io/projected/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-kube-api-access-hzfdm\") pod \"placement-db-sync-459j5\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.635698 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cndl\" (UniqueName: \"kubernetes.io/projected/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-kube-api-access-2cndl\") pod \"horizon-7d5695b9d5-jnxdq\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667629 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667669 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-scripts\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667712 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667760 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-log-httpd\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667783 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-run-httpd\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667818 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-config\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667841 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpdhr\" (UniqueName: \"kubernetes.io/projected/89017745-df9d-4bae-ad06-786a6b653ce8-kube-api-access-wpdhr\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667878 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667894 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-config-data\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667916 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667936 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.667963 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv57g\" (UniqueName: \"kubernetes.io/projected/74dc644a-498d-47b5-af2d-164d5eee997a-kube-api-access-tv57g\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.668027 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.673862 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-log-httpd\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.676343 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-run-httpd\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.685142 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.685381 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-config-data\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.687334 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.701919 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-scripts\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.711274 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpdhr\" (UniqueName: \"kubernetes.io/projected/89017745-df9d-4bae-ad06-786a6b653ce8-kube-api-access-wpdhr\") pod \"ceilometer-0\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.770133 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-config\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.770208 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.770263 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.770316 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv57g\" (UniqueName: \"kubernetes.io/projected/74dc644a-498d-47b5-af2d-164d5eee997a-kube-api-access-tv57g\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.770413 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.770450 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.772330 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-config\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.775084 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.785855 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.786269 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.789685 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6qj89" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.798358 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.843849 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.845337 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv57g\" (UniqueName: \"kubernetes.io/projected/74dc644a-498d-47b5-af2d-164d5eee997a-kube-api-access-tv57g\") pod \"dnsmasq-dns-fcfdd6f9f-cwbzl\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.888487 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-459j5" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.942928 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.952703 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:36 crc kubenswrapper[4575]: I1004 04:55:36.984628 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nqbf7"] Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.048282 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-dkfg2"] Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.307617 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d7cd44969-cjg5j"] Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.377915 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12365747-76ba-4bcb-a828-147712279d49" path="/var/lib/kubelet/pods/12365747-76ba-4bcb-a828-147712279d49/volumes" Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.481950 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" event={"ID":"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93","Type":"ContainerStarted","Data":"0018b21d0cf187e7029ccb3097632af09b091d32baa1d5c2423e4cba19af93c4"} Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.500464 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d7cd44969-cjg5j" event={"ID":"27553b01-0eaa-4cad-96f2-afca79cbefc5","Type":"ContainerStarted","Data":"0ed0050a4867cfbd3a32a7efc1bb8069619a7a30188280dd079031f6b87f3410"} Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.524996 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nqbf7" event={"ID":"0584c5bf-98ef-462d-b62d-8211acd79e8e","Type":"ContainerStarted","Data":"cd82951ce135d36bc283075a3be6582925b7dcd041877fa199218305c62ddc17"} Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.575523 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kg5nx"] Oct 04 04:55:37 crc kubenswrapper[4575]: W1004 04:55:37.604573 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef74f3cc_9958_4de3_a914_12e5453e169d.slice/crio-f82c00dce2184a88862328a6fb828aeaa8930d1322e65c3b5e293110beb79f3f WatchSource:0}: Error finding container f82c00dce2184a88862328a6fb828aeaa8930d1322e65c3b5e293110beb79f3f: Status 404 returned error can't find the container with id f82c00dce2184a88862328a6fb828aeaa8930d1322e65c3b5e293110beb79f3f Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.616430 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d5695b9d5-jnxdq"] Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.821640 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-459j5"] Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.841490 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6qj89"] Oct 04 04:55:37 crc kubenswrapper[4575]: W1004 04:55:37.855759 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62f6c233_faeb_47e1_8103_4902eaea8d4a.slice/crio-ae4e82bea55252f6efc08bc517eb234bfb8319ff1e23a1d674a03ac36c9026a8 WatchSource:0}: Error finding container ae4e82bea55252f6efc08bc517eb234bfb8319ff1e23a1d674a03ac36c9026a8: Status 404 returned error can't find the container with id ae4e82bea55252f6efc08bc517eb234bfb8319ff1e23a1d674a03ac36c9026a8 Oct 04 04:55:37 crc kubenswrapper[4575]: I1004 04:55:37.971823 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-cwbzl"] Oct 04 04:55:38 crc kubenswrapper[4575]: W1004 04:55:38.022046 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89017745_df9d_4bae_ad06_786a6b653ce8.slice/crio-dbb201289bea12734054ef5d370e316daf3fb6c76bda22f46d54827d6792995b WatchSource:0}: Error finding container dbb201289bea12734054ef5d370e316daf3fb6c76bda22f46d54827d6792995b: Status 404 returned error can't find the container with id dbb201289bea12734054ef5d370e316daf3fb6c76bda22f46d54827d6792995b Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.031308 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.543924 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6qj89" event={"ID":"62f6c233-faeb-47e1-8103-4902eaea8d4a","Type":"ContainerStarted","Data":"ae4e82bea55252f6efc08bc517eb234bfb8319ff1e23a1d674a03ac36c9026a8"} Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.556919 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89017745-df9d-4bae-ad06-786a6b653ce8","Type":"ContainerStarted","Data":"dbb201289bea12734054ef5d370e316daf3fb6c76bda22f46d54827d6792995b"} Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.580696 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d5695b9d5-jnxdq" event={"ID":"32ffd4c7-c1dc-4d05-8498-5f496d52c40e","Type":"ContainerStarted","Data":"38226b82e62aae441a184da49a8ea02642cbdccc877e570d15858985af1e2c9b"} Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.604009 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nqbf7" event={"ID":"0584c5bf-98ef-462d-b62d-8211acd79e8e","Type":"ContainerStarted","Data":"63ea1872a641b4a425e2746229cfdec26f0609d439e87d391c014c1d4a2355b0"} Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.637080 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nqbf7" podStartSLOduration=3.637053524 podStartE2EDuration="3.637053524s" podCreationTimestamp="2025-10-04 04:55:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:38.632769051 +0000 UTC m=+1289.961327865" watchObservedRunningTime="2025-10-04 04:55:38.637053524 +0000 UTC m=+1289.965612358" Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.659978 4575 generic.go:334] "Generic (PLEG): container finished" podID="93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" containerID="9f3fc8ab7a413e0c106ca89a5d9a185f2a1750cbb9f84c228c094066544e413c" exitCode=0 Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.660133 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" event={"ID":"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93","Type":"ContainerDied","Data":"9f3fc8ab7a413e0c106ca89a5d9a185f2a1750cbb9f84c228c094066544e413c"} Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.669027 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kg5nx" event={"ID":"ef74f3cc-9958-4de3-a914-12e5453e169d","Type":"ContainerStarted","Data":"f82c00dce2184a88862328a6fb828aeaa8930d1322e65c3b5e293110beb79f3f"} Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.697885 4575 generic.go:334] "Generic (PLEG): container finished" podID="74dc644a-498d-47b5-af2d-164d5eee997a" containerID="7f114fbc003f7b8d37daff943283cc53ee1d8a75880f39dfd86d976c49165342" exitCode=0 Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.698014 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" event={"ID":"74dc644a-498d-47b5-af2d-164d5eee997a","Type":"ContainerDied","Data":"7f114fbc003f7b8d37daff943283cc53ee1d8a75880f39dfd86d976c49165342"} Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.698045 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" event={"ID":"74dc644a-498d-47b5-af2d-164d5eee997a","Type":"ContainerStarted","Data":"fc6790f29ffd6bbd941e75b6bc321187af2b6ab61ec7dd0ea7f44a1ec5dcc31a"} Oct 04 04:55:38 crc kubenswrapper[4575]: I1004 04:55:38.720212 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-459j5" event={"ID":"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92","Type":"ContainerStarted","Data":"1dda133b2d3dcc676eac03d1beb1ed0afc02a66e634dc4304c63f748a7ad338c"} Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.076850 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d7cd44969-cjg5j"] Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.150315 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-647567c5b9-xst9n"] Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.152260 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.180146 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-647567c5b9-xst9n"] Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.201038 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.282959 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-scripts\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.283067 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw4sz\" (UniqueName: \"kubernetes.io/projected/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-kube-api-access-pw4sz\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.283116 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-logs\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.283164 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-horizon-secret-key\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.283198 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-config-data\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.385079 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-config-data\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.386336 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-config-data\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.387130 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-scripts\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.389938 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw4sz\" (UniqueName: \"kubernetes.io/projected/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-kube-api-access-pw4sz\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.390615 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-logs\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.390824 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-horizon-secret-key\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.391003 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-scripts\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.391543 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-logs\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.439078 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw4sz\" (UniqueName: \"kubernetes.io/projected/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-kube-api-access-pw4sz\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.480637 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-horizon-secret-key\") pod \"horizon-647567c5b9-xst9n\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.488181 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.538723 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.597396 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-config\") pod \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.597825 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-nb\") pod \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.597925 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-swift-storage-0\") pod \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.598188 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-svc\") pod \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.598275 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbt7r\" (UniqueName: \"kubernetes.io/projected/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-kube-api-access-xbt7r\") pod \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.598319 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-sb\") pod \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\" (UID: \"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93\") " Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.634148 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-kube-api-access-xbt7r" (OuterVolumeSpecName: "kube-api-access-xbt7r") pod "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" (UID: "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93"). InnerVolumeSpecName "kube-api-access-xbt7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.655865 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" (UID: "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.684808 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" (UID: "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.698687 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" (UID: "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.699982 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbt7r\" (UniqueName: \"kubernetes.io/projected/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-kube-api-access-xbt7r\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.700017 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.700030 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.700045 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.710157 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" (UID: "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.721482 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-config" (OuterVolumeSpecName: "config") pod "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" (UID: "93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.784239 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.784552 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-dkfg2" event={"ID":"93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93","Type":"ContainerDied","Data":"0018b21d0cf187e7029ccb3097632af09b091d32baa1d5c2423e4cba19af93c4"} Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.784661 4575 scope.go:117] "RemoveContainer" containerID="9f3fc8ab7a413e0c106ca89a5d9a185f2a1750cbb9f84c228c094066544e413c" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.805558 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" event={"ID":"74dc644a-498d-47b5-af2d-164d5eee997a","Type":"ContainerStarted","Data":"08c4ccb92fab96b41417b0a4ae95f796b798dee566942cf7a1a9aea9ea4c94b5"} Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.805651 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.805789 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.805818 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.847242 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" podStartSLOduration=3.847225544 podStartE2EDuration="3.847225544s" podCreationTimestamp="2025-10-04 04:55:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:39.833223221 +0000 UTC m=+1291.161782045" watchObservedRunningTime="2025-10-04 04:55:39.847225544 +0000 UTC m=+1291.175784348" Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.927728 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-dkfg2"] Oct 04 04:55:39 crc kubenswrapper[4575]: I1004 04:55:39.941781 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-dkfg2"] Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.403783 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-647567c5b9-xst9n"] Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.493742 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-7kr5n"] Oct 04 04:55:40 crc kubenswrapper[4575]: E1004 04:55:40.494299 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" containerName="init" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.494322 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" containerName="init" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.494539 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" containerName="init" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.495252 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.500687 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rxnmx" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.500987 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.501606 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.532825 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-7kr5n"] Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.631802 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-272pp\" (UniqueName: \"kubernetes.io/projected/d089c375-6e05-4d74-88e1-d00c8cbddb1f-kube-api-access-272pp\") pod \"neutron-db-sync-7kr5n\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.631891 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-config\") pod \"neutron-db-sync-7kr5n\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.631921 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-combined-ca-bundle\") pod \"neutron-db-sync-7kr5n\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.736018 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-config\") pod \"neutron-db-sync-7kr5n\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.736077 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-combined-ca-bundle\") pod \"neutron-db-sync-7kr5n\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.736254 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-272pp\" (UniqueName: \"kubernetes.io/projected/d089c375-6e05-4d74-88e1-d00c8cbddb1f-kube-api-access-272pp\") pod \"neutron-db-sync-7kr5n\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.771998 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-combined-ca-bundle\") pod \"neutron-db-sync-7kr5n\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.774833 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-272pp\" (UniqueName: \"kubernetes.io/projected/d089c375-6e05-4d74-88e1-d00c8cbddb1f-kube-api-access-272pp\") pod \"neutron-db-sync-7kr5n\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.800369 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-config\") pod \"neutron-db-sync-7kr5n\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.827023 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:55:40 crc kubenswrapper[4575]: I1004 04:55:40.831999 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-647567c5b9-xst9n" event={"ID":"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9","Type":"ContainerStarted","Data":"93767a39f1512a8289ade025feaebc4a9c81b71bd905a700465394e7e4940074"} Oct 04 04:55:41 crc kubenswrapper[4575]: I1004 04:55:41.328738 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93" path="/var/lib/kubelet/pods/93a1746e-ea0e-4bf3-a4ed-1ac29daa3d93/volumes" Oct 04 04:55:41 crc kubenswrapper[4575]: I1004 04:55:41.427740 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-7kr5n"] Oct 04 04:55:41 crc kubenswrapper[4575]: I1004 04:55:41.863676 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-7kr5n" event={"ID":"d089c375-6e05-4d74-88e1-d00c8cbddb1f","Type":"ContainerStarted","Data":"3bd213d96e5e86bf506ad4db79bad1f7fde3805d418daa4d347cfcd31d0ed19e"} Oct 04 04:55:41 crc kubenswrapper[4575]: I1004 04:55:41.864054 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-7kr5n" event={"ID":"d089c375-6e05-4d74-88e1-d00c8cbddb1f","Type":"ContainerStarted","Data":"c06b6d57dbeae7f85c9ccf06f07cc4af6fa0cbb44fce0b2bda78427bd0245a7b"} Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.171323 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-7kr5n" podStartSLOduration=5.171301528 podStartE2EDuration="5.171301528s" podCreationTimestamp="2025-10-04 04:55:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:42.927573913 +0000 UTC m=+1294.256132727" watchObservedRunningTime="2025-10-04 04:55:45.171301528 +0000 UTC m=+1296.499860332" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.177130 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d5695b9d5-jnxdq"] Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.221108 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-b86b879b4-tlf42"] Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.223144 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.229423 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.254166 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b86b879b4-tlf42"] Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.314555 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-scripts\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.314620 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-secret-key\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.314670 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4rf6\" (UniqueName: \"kubernetes.io/projected/115cafce-0f37-4b82-8b55-c2bbf7297226-kube-api-access-v4rf6\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.314707 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-tls-certs\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.314738 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/115cafce-0f37-4b82-8b55-c2bbf7297226-logs\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.314771 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-config-data\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.314796 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-combined-ca-bundle\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.358402 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-647567c5b9-xst9n"] Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.387368 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-644bf5cdd4-pwdw9"] Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.390656 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.412340 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-644bf5cdd4-pwdw9"] Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.424979 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-scripts\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.425066 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-secret-key\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.425140 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4rf6\" (UniqueName: \"kubernetes.io/projected/115cafce-0f37-4b82-8b55-c2bbf7297226-kube-api-access-v4rf6\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.425211 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-tls-certs\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.425291 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/115cafce-0f37-4b82-8b55-c2bbf7297226-logs\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.425398 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-config-data\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.425461 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-combined-ca-bundle\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.426212 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/115cafce-0f37-4b82-8b55-c2bbf7297226-logs\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.426306 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-scripts\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.438258 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-combined-ca-bundle\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.439391 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-config-data\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.440771 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-secret-key\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.460663 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-tls-certs\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.480345 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4rf6\" (UniqueName: \"kubernetes.io/projected/115cafce-0f37-4b82-8b55-c2bbf7297226-kube-api-access-v4rf6\") pod \"horizon-b86b879b4-tlf42\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.542363 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f0573f0-1418-43eb-be82-166fb99ed23d-combined-ca-bundle\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.542465 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg7x9\" (UniqueName: \"kubernetes.io/projected/6f0573f0-1418-43eb-be82-166fb99ed23d-kube-api-access-tg7x9\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.542493 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6f0573f0-1418-43eb-be82-166fb99ed23d-horizon-secret-key\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.542513 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f0573f0-1418-43eb-be82-166fb99ed23d-horizon-tls-certs\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.542553 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f0573f0-1418-43eb-be82-166fb99ed23d-scripts\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.542577 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f0573f0-1418-43eb-be82-166fb99ed23d-config-data\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.542631 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f0573f0-1418-43eb-be82-166fb99ed23d-logs\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.556608 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.643778 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg7x9\" (UniqueName: \"kubernetes.io/projected/6f0573f0-1418-43eb-be82-166fb99ed23d-kube-api-access-tg7x9\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.643845 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6f0573f0-1418-43eb-be82-166fb99ed23d-horizon-secret-key\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.643865 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f0573f0-1418-43eb-be82-166fb99ed23d-horizon-tls-certs\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.643903 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f0573f0-1418-43eb-be82-166fb99ed23d-scripts\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.643920 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f0573f0-1418-43eb-be82-166fb99ed23d-config-data\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.643945 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f0573f0-1418-43eb-be82-166fb99ed23d-logs\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.644033 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f0573f0-1418-43eb-be82-166fb99ed23d-combined-ca-bundle\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.645120 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f0573f0-1418-43eb-be82-166fb99ed23d-logs\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.646063 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f0573f0-1418-43eb-be82-166fb99ed23d-scripts\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.649053 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6f0573f0-1418-43eb-be82-166fb99ed23d-horizon-secret-key\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.651256 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f0573f0-1418-43eb-be82-166fb99ed23d-config-data\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.665536 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg7x9\" (UniqueName: \"kubernetes.io/projected/6f0573f0-1418-43eb-be82-166fb99ed23d-kube-api-access-tg7x9\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.667393 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f0573f0-1418-43eb-be82-166fb99ed23d-horizon-tls-certs\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.670639 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f0573f0-1418-43eb-be82-166fb99ed23d-combined-ca-bundle\") pod \"horizon-644bf5cdd4-pwdw9\" (UID: \"6f0573f0-1418-43eb-be82-166fb99ed23d\") " pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:45 crc kubenswrapper[4575]: I1004 04:55:45.713213 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:55:46 crc kubenswrapper[4575]: I1004 04:55:46.959257 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:55:47 crc kubenswrapper[4575]: I1004 04:55:47.071992 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-ttlxx"] Oct 04 04:55:47 crc kubenswrapper[4575]: I1004 04:55:47.072567 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="dnsmasq-dns" containerID="cri-o://da60a53726dd43ff90d13c3b3f93974566e2f3b78dfb1904ec524d845f90485e" gracePeriod=10 Oct 04 04:55:47 crc kubenswrapper[4575]: I1004 04:55:47.740223 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Oct 04 04:55:47 crc kubenswrapper[4575]: I1004 04:55:47.951911 4575 generic.go:334] "Generic (PLEG): container finished" podID="6dca86a7-57a1-457e-adf9-58882540f233" containerID="da60a53726dd43ff90d13c3b3f93974566e2f3b78dfb1904ec524d845f90485e" exitCode=0 Oct 04 04:55:47 crc kubenswrapper[4575]: I1004 04:55:47.951976 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" event={"ID":"6dca86a7-57a1-457e-adf9-58882540f233","Type":"ContainerDied","Data":"da60a53726dd43ff90d13c3b3f93974566e2f3b78dfb1904ec524d845f90485e"} Oct 04 04:55:47 crc kubenswrapper[4575]: I1004 04:55:47.953882 4575 generic.go:334] "Generic (PLEG): container finished" podID="0584c5bf-98ef-462d-b62d-8211acd79e8e" containerID="63ea1872a641b4a425e2746229cfdec26f0609d439e87d391c014c1d4a2355b0" exitCode=0 Oct 04 04:55:47 crc kubenswrapper[4575]: I1004 04:55:47.953940 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nqbf7" event={"ID":"0584c5bf-98ef-462d-b62d-8211acd79e8e","Type":"ContainerDied","Data":"63ea1872a641b4a425e2746229cfdec26f0609d439e87d391c014c1d4a2355b0"} Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.169754 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.315248 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-credential-keys\") pod \"0584c5bf-98ef-462d-b62d-8211acd79e8e\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.315315 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxp57\" (UniqueName: \"kubernetes.io/projected/0584c5bf-98ef-462d-b62d-8211acd79e8e-kube-api-access-zxp57\") pod \"0584c5bf-98ef-462d-b62d-8211acd79e8e\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.315394 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-fernet-keys\") pod \"0584c5bf-98ef-462d-b62d-8211acd79e8e\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.315476 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-scripts\") pod \"0584c5bf-98ef-462d-b62d-8211acd79e8e\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.315519 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-config-data\") pod \"0584c5bf-98ef-462d-b62d-8211acd79e8e\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.315552 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-combined-ca-bundle\") pod \"0584c5bf-98ef-462d-b62d-8211acd79e8e\" (UID: \"0584c5bf-98ef-462d-b62d-8211acd79e8e\") " Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.322355 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0584c5bf-98ef-462d-b62d-8211acd79e8e" (UID: "0584c5bf-98ef-462d-b62d-8211acd79e8e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.323008 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0584c5bf-98ef-462d-b62d-8211acd79e8e" (UID: "0584c5bf-98ef-462d-b62d-8211acd79e8e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.324955 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-scripts" (OuterVolumeSpecName: "scripts") pod "0584c5bf-98ef-462d-b62d-8211acd79e8e" (UID: "0584c5bf-98ef-462d-b62d-8211acd79e8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.329949 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0584c5bf-98ef-462d-b62d-8211acd79e8e-kube-api-access-zxp57" (OuterVolumeSpecName: "kube-api-access-zxp57") pod "0584c5bf-98ef-462d-b62d-8211acd79e8e" (UID: "0584c5bf-98ef-462d-b62d-8211acd79e8e"). InnerVolumeSpecName "kube-api-access-zxp57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.349865 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-config-data" (OuterVolumeSpecName: "config-data") pod "0584c5bf-98ef-462d-b62d-8211acd79e8e" (UID: "0584c5bf-98ef-462d-b62d-8211acd79e8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.364408 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0584c5bf-98ef-462d-b62d-8211acd79e8e" (UID: "0584c5bf-98ef-462d-b62d-8211acd79e8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.419153 4575 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.419227 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.419242 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.419255 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.419271 4575 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0584c5bf-98ef-462d-b62d-8211acd79e8e-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4575]: I1004 04:55:52.419283 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxp57\" (UniqueName: \"kubernetes.io/projected/0584c5bf-98ef-462d-b62d-8211acd79e8e-kube-api-access-zxp57\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.008262 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nqbf7" event={"ID":"0584c5bf-98ef-462d-b62d-8211acd79e8e","Type":"ContainerDied","Data":"cd82951ce135d36bc283075a3be6582925b7dcd041877fa199218305c62ddc17"} Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.008646 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd82951ce135d36bc283075a3be6582925b7dcd041877fa199218305c62ddc17" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.008360 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nqbf7" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.260259 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nqbf7"] Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.268938 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nqbf7"] Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.322556 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0584c5bf-98ef-462d-b62d-8211acd79e8e" path="/var/lib/kubelet/pods/0584c5bf-98ef-462d-b62d-8211acd79e8e/volumes" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.356955 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-99bks"] Oct 04 04:55:53 crc kubenswrapper[4575]: E1004 04:55:53.357343 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0584c5bf-98ef-462d-b62d-8211acd79e8e" containerName="keystone-bootstrap" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.357366 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="0584c5bf-98ef-462d-b62d-8211acd79e8e" containerName="keystone-bootstrap" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.357536 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="0584c5bf-98ef-462d-b62d-8211acd79e8e" containerName="keystone-bootstrap" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.358271 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.361860 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.361995 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.362352 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xdgq4" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.362467 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.373760 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-99bks"] Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.538457 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-config-data\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.538508 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvll6\" (UniqueName: \"kubernetes.io/projected/ca934f38-b0b0-4d26-bf2e-63f983c81851-kube-api-access-dvll6\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.538552 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-scripts\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.538574 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-fernet-keys\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.538647 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-credential-keys\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.538673 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-combined-ca-bundle\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.640078 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-config-data\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.640124 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvll6\" (UniqueName: \"kubernetes.io/projected/ca934f38-b0b0-4d26-bf2e-63f983c81851-kube-api-access-dvll6\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.640154 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-scripts\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.640173 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-fernet-keys\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.640219 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-credential-keys\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.640239 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-combined-ca-bundle\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.644516 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-config-data\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.644819 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-scripts\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.645403 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-combined-ca-bundle\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.646524 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-fernet-keys\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.647260 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-credential-keys\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.659894 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvll6\" (UniqueName: \"kubernetes.io/projected/ca934f38-b0b0-4d26-bf2e-63f983c81851-kube-api-access-dvll6\") pod \"keystone-bootstrap-99bks\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: I1004 04:55:53.687311 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-99bks" Oct 04 04:55:53 crc kubenswrapper[4575]: E1004 04:55:53.698929 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 04 04:55:53 crc kubenswrapper[4575]: E1004 04:55:53.699078 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hfdtp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-6qj89_openstack(62f6c233-faeb-47e1-8103-4902eaea8d4a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:55:53 crc kubenswrapper[4575]: E1004 04:55:53.700467 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-6qj89" podUID="62f6c233-faeb-47e1-8103-4902eaea8d4a" Oct 04 04:55:54 crc kubenswrapper[4575]: E1004 04:55:54.030660 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-6qj89" podUID="62f6c233-faeb-47e1-8103-4902eaea8d4a" Oct 04 04:55:57 crc kubenswrapper[4575]: I1004 04:55:57.740536 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Oct 04 04:56:01 crc kubenswrapper[4575]: E1004 04:56:01.233174 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 04:56:01 crc kubenswrapper[4575]: E1004 04:56:01.234012 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n647h5d6h5cch58fh6bh545h85h5c9h7ch545h4h5dhcch95h589h668h5b6h5c8h647h58fh685hf4h54fh58dh5d5h574h54h57bhffh64h8fh658q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2cndl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7d5695b9d5-jnxdq_openstack(32ffd4c7-c1dc-4d05-8498-5f496d52c40e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:56:01 crc kubenswrapper[4575]: E1004 04:56:01.238405 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7d5695b9d5-jnxdq" podUID="32ffd4c7-c1dc-4d05-8498-5f496d52c40e" Oct 04 04:56:01 crc kubenswrapper[4575]: E1004 04:56:01.273429 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 04:56:01 crc kubenswrapper[4575]: E1004 04:56:01.273674 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n67chf5hb9h59chf7h98hcbh5cbh5fbh5d5h59chdfh66ch78h67dh58h697h86h5dfh5f4h597h587hddhb9h54bh649h686h557h5ch79h5b6h64cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zwtlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7d7cd44969-cjg5j_openstack(27553b01-0eaa-4cad-96f2-afca79cbefc5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:56:01 crc kubenswrapper[4575]: E1004 04:56:01.277172 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7d7cd44969-cjg5j" podUID="27553b01-0eaa-4cad-96f2-afca79cbefc5" Oct 04 04:56:02 crc kubenswrapper[4575]: I1004 04:56:02.741745 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Oct 04 04:56:02 crc kubenswrapper[4575]: I1004 04:56:02.742863 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.114004 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" event={"ID":"6dca86a7-57a1-457e-adf9-58882540f233","Type":"ContainerDied","Data":"ef2c4f1f74dd6cd838724c79409e3d78d6f01ecf7ff654427bea017764bf3041"} Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.114038 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef2c4f1f74dd6cd838724c79409e3d78d6f01ecf7ff654427bea017764bf3041" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.186772 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.317644 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsdnb\" (UniqueName: \"kubernetes.io/projected/6dca86a7-57a1-457e-adf9-58882540f233-kube-api-access-vsdnb\") pod \"6dca86a7-57a1-457e-adf9-58882540f233\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.317740 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-nb\") pod \"6dca86a7-57a1-457e-adf9-58882540f233\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.317781 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-config\") pod \"6dca86a7-57a1-457e-adf9-58882540f233\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.318034 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-swift-storage-0\") pod \"6dca86a7-57a1-457e-adf9-58882540f233\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.318080 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-svc\") pod \"6dca86a7-57a1-457e-adf9-58882540f233\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.318253 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-sb\") pod \"6dca86a7-57a1-457e-adf9-58882540f233\" (UID: \"6dca86a7-57a1-457e-adf9-58882540f233\") " Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.341434 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dca86a7-57a1-457e-adf9-58882540f233-kube-api-access-vsdnb" (OuterVolumeSpecName: "kube-api-access-vsdnb") pod "6dca86a7-57a1-457e-adf9-58882540f233" (UID: "6dca86a7-57a1-457e-adf9-58882540f233"). InnerVolumeSpecName "kube-api-access-vsdnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.367845 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6dca86a7-57a1-457e-adf9-58882540f233" (UID: "6dca86a7-57a1-457e-adf9-58882540f233"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.369040 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6dca86a7-57a1-457e-adf9-58882540f233" (UID: "6dca86a7-57a1-457e-adf9-58882540f233"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.370662 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-config" (OuterVolumeSpecName: "config") pod "6dca86a7-57a1-457e-adf9-58882540f233" (UID: "6dca86a7-57a1-457e-adf9-58882540f233"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.381293 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6dca86a7-57a1-457e-adf9-58882540f233" (UID: "6dca86a7-57a1-457e-adf9-58882540f233"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.388929 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6dca86a7-57a1-457e-adf9-58882540f233" (UID: "6dca86a7-57a1-457e-adf9-58882540f233"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.421767 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsdnb\" (UniqueName: \"kubernetes.io/projected/6dca86a7-57a1-457e-adf9-58882540f233-kube-api-access-vsdnb\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.421811 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.421825 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.421840 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.421853 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:03 crc kubenswrapper[4575]: I1004 04:56:03.421901 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dca86a7-57a1-457e-adf9-58882540f233-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:04 crc kubenswrapper[4575]: I1004 04:56:04.133368 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" Oct 04 04:56:04 crc kubenswrapper[4575]: I1004 04:56:04.174461 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-ttlxx"] Oct 04 04:56:04 crc kubenswrapper[4575]: I1004 04:56:04.183316 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-ttlxx"] Oct 04 04:56:05 crc kubenswrapper[4575]: I1004 04:56:05.323708 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dca86a7-57a1-457e-adf9-58882540f233" path="/var/lib/kubelet/pods/6dca86a7-57a1-457e-adf9-58882540f233/volumes" Oct 04 04:56:07 crc kubenswrapper[4575]: I1004 04:56:07.743433 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-ttlxx" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Oct 04 04:56:24 crc kubenswrapper[4575]: E1004 04:56:24.501829 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 04:56:24 crc kubenswrapper[4575]: E1004 04:56:24.502439 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n568hbch94h64bh655h7dhf7h584h76h699h5bbh68dhfdh78hdfh64dh695h659h65dh5d4hb9h589h5bbhb4h87h94h67bh675h694hddh74h597q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pw4sz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-647567c5b9-xst9n_openstack(1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:56:24 crc kubenswrapper[4575]: E1004 04:56:24.505531 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-647567c5b9-xst9n" podUID="1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.639919 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.645953 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.721944 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-scripts\") pod \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722000 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cndl\" (UniqueName: \"kubernetes.io/projected/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-kube-api-access-2cndl\") pod \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722020 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-config-data\") pod \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722046 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-config-data\") pod \"27553b01-0eaa-4cad-96f2-afca79cbefc5\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722077 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-logs\") pod \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722097 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27553b01-0eaa-4cad-96f2-afca79cbefc5-logs\") pod \"27553b01-0eaa-4cad-96f2-afca79cbefc5\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722123 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-horizon-secret-key\") pod \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\" (UID: \"32ffd4c7-c1dc-4d05-8498-5f496d52c40e\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722183 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-scripts\") pod \"27553b01-0eaa-4cad-96f2-afca79cbefc5\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722237 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwtlp\" (UniqueName: \"kubernetes.io/projected/27553b01-0eaa-4cad-96f2-afca79cbefc5-kube-api-access-zwtlp\") pod \"27553b01-0eaa-4cad-96f2-afca79cbefc5\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722265 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27553b01-0eaa-4cad-96f2-afca79cbefc5-horizon-secret-key\") pod \"27553b01-0eaa-4cad-96f2-afca79cbefc5\" (UID: \"27553b01-0eaa-4cad-96f2-afca79cbefc5\") " Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722835 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-logs" (OuterVolumeSpecName: "logs") pod "32ffd4c7-c1dc-4d05-8498-5f496d52c40e" (UID: "32ffd4c7-c1dc-4d05-8498-5f496d52c40e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.722908 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-scripts" (OuterVolumeSpecName: "scripts") pod "32ffd4c7-c1dc-4d05-8498-5f496d52c40e" (UID: "32ffd4c7-c1dc-4d05-8498-5f496d52c40e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.723167 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27553b01-0eaa-4cad-96f2-afca79cbefc5-logs" (OuterVolumeSpecName: "logs") pod "27553b01-0eaa-4cad-96f2-afca79cbefc5" (UID: "27553b01-0eaa-4cad-96f2-afca79cbefc5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.723478 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-scripts" (OuterVolumeSpecName: "scripts") pod "27553b01-0eaa-4cad-96f2-afca79cbefc5" (UID: "27553b01-0eaa-4cad-96f2-afca79cbefc5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.723944 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-config-data" (OuterVolumeSpecName: "config-data") pod "32ffd4c7-c1dc-4d05-8498-5f496d52c40e" (UID: "32ffd4c7-c1dc-4d05-8498-5f496d52c40e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.728607 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.728626 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.728635 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.728642 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27553b01-0eaa-4cad-96f2-afca79cbefc5-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.728650 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.728983 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-config-data" (OuterVolumeSpecName: "config-data") pod "27553b01-0eaa-4cad-96f2-afca79cbefc5" (UID: "27553b01-0eaa-4cad-96f2-afca79cbefc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.749839 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27553b01-0eaa-4cad-96f2-afca79cbefc5-kube-api-access-zwtlp" (OuterVolumeSpecName: "kube-api-access-zwtlp") pod "27553b01-0eaa-4cad-96f2-afca79cbefc5" (UID: "27553b01-0eaa-4cad-96f2-afca79cbefc5"). InnerVolumeSpecName "kube-api-access-zwtlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.750202 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "32ffd4c7-c1dc-4d05-8498-5f496d52c40e" (UID: "32ffd4c7-c1dc-4d05-8498-5f496d52c40e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.750477 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-kube-api-access-2cndl" (OuterVolumeSpecName: "kube-api-access-2cndl") pod "32ffd4c7-c1dc-4d05-8498-5f496d52c40e" (UID: "32ffd4c7-c1dc-4d05-8498-5f496d52c40e"). InnerVolumeSpecName "kube-api-access-2cndl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.753353 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27553b01-0eaa-4cad-96f2-afca79cbefc5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "27553b01-0eaa-4cad-96f2-afca79cbefc5" (UID: "27553b01-0eaa-4cad-96f2-afca79cbefc5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.829791 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cndl\" (UniqueName: \"kubernetes.io/projected/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-kube-api-access-2cndl\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.829824 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27553b01-0eaa-4cad-96f2-afca79cbefc5-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.829836 4575 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32ffd4c7-c1dc-4d05-8498-5f496d52c40e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.829847 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwtlp\" (UniqueName: \"kubernetes.io/projected/27553b01-0eaa-4cad-96f2-afca79cbefc5-kube-api-access-zwtlp\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:24 crc kubenswrapper[4575]: I1004 04:56:24.829857 4575 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27553b01-0eaa-4cad-96f2-afca79cbefc5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:25 crc kubenswrapper[4575]: I1004 04:56:25.008889 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b86b879b4-tlf42"] Oct 04 04:56:25 crc kubenswrapper[4575]: E1004 04:56:25.096317 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 04 04:56:25 crc kubenswrapper[4575]: E1004 04:56:25.096498 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bch56dh5fhc8h96hdbh698h656h545h5dbh5bch64ch596hc9h8fh64chdch56h679h54fh558h686h649h5b9h588h58h58bh7bh68h76h584h695q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wpdhr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(89017745-df9d-4bae-ad06-786a6b653ce8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:56:25 crc kubenswrapper[4575]: I1004 04:56:25.339892 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d7cd44969-cjg5j" Oct 04 04:56:25 crc kubenswrapper[4575]: I1004 04:56:25.339909 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d7cd44969-cjg5j" event={"ID":"27553b01-0eaa-4cad-96f2-afca79cbefc5","Type":"ContainerDied","Data":"0ed0050a4867cfbd3a32a7efc1bb8069619a7a30188280dd079031f6b87f3410"} Oct 04 04:56:25 crc kubenswrapper[4575]: I1004 04:56:25.345179 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d5695b9d5-jnxdq" event={"ID":"32ffd4c7-c1dc-4d05-8498-5f496d52c40e","Type":"ContainerDied","Data":"38226b82e62aae441a184da49a8ea02642cbdccc877e570d15858985af1e2c9b"} Oct 04 04:56:25 crc kubenswrapper[4575]: I1004 04:56:25.345216 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d5695b9d5-jnxdq" Oct 04 04:56:25 crc kubenswrapper[4575]: I1004 04:56:25.463291 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d7cd44969-cjg5j"] Oct 04 04:56:25 crc kubenswrapper[4575]: I1004 04:56:25.488087 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d7cd44969-cjg5j"] Oct 04 04:56:25 crc kubenswrapper[4575]: I1004 04:56:25.516165 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d5695b9d5-jnxdq"] Oct 04 04:56:25 crc kubenswrapper[4575]: I1004 04:56:25.525753 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d5695b9d5-jnxdq"] Oct 04 04:56:26 crc kubenswrapper[4575]: I1004 04:56:26.357089 4575 generic.go:334] "Generic (PLEG): container finished" podID="775d96cb-5e3b-437f-b61d-7799e86f2f58" containerID="b88c3e2d3ac777ec98e5b700d75c1cfaf4895dd17819590c14bf1ffc0030ec62" exitCode=0 Oct 04 04:56:26 crc kubenswrapper[4575]: I1004 04:56:26.357172 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gtvfg" event={"ID":"775d96cb-5e3b-437f-b61d-7799e86f2f58","Type":"ContainerDied","Data":"b88c3e2d3ac777ec98e5b700d75c1cfaf4895dd17819590c14bf1ffc0030ec62"} Oct 04 04:56:26 crc kubenswrapper[4575]: E1004 04:56:26.810684 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 04 04:56:26 crc kubenswrapper[4575]: E1004 04:56:26.811363 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2tjm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-kg5nx_openstack(ef74f3cc-9958-4de3-a914-12e5453e169d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:56:26 crc kubenswrapper[4575]: E1004 04:56:26.813844 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-kg5nx" podUID="ef74f3cc-9958-4de3-a914-12e5453e169d" Oct 04 04:56:26 crc kubenswrapper[4575]: I1004 04:56:26.936848 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.069859 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-config-data\") pod \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.070314 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-horizon-secret-key\") pod \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.070356 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-logs\") pod \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.070418 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw4sz\" (UniqueName: \"kubernetes.io/projected/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-kube-api-access-pw4sz\") pod \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.070459 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-scripts\") pod \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\" (UID: \"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9\") " Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.070804 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-logs" (OuterVolumeSpecName: "logs") pod "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9" (UID: "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.070968 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.071277 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-scripts" (OuterVolumeSpecName: "scripts") pod "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9" (UID: "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.071327 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-config-data" (OuterVolumeSpecName: "config-data") pod "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9" (UID: "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.074163 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-kube-api-access-pw4sz" (OuterVolumeSpecName: "kube-api-access-pw4sz") pod "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9" (UID: "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9"). InnerVolumeSpecName "kube-api-access-pw4sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.075219 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9" (UID: "1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.176700 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw4sz\" (UniqueName: \"kubernetes.io/projected/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-kube-api-access-pw4sz\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.176740 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.176750 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.176761 4575 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.274457 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-644bf5cdd4-pwdw9"] Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.322649 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27553b01-0eaa-4cad-96f2-afca79cbefc5" path="/var/lib/kubelet/pods/27553b01-0eaa-4cad-96f2-afca79cbefc5/volumes" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.323252 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32ffd4c7-c1dc-4d05-8498-5f496d52c40e" path="/var/lib/kubelet/pods/32ffd4c7-c1dc-4d05-8498-5f496d52c40e/volumes" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.381568 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-99bks"] Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.387068 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-647567c5b9-xst9n" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.387063 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-647567c5b9-xst9n" event={"ID":"1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9","Type":"ContainerDied","Data":"93767a39f1512a8289ade025feaebc4a9c81b71bd905a700465394e7e4940074"} Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.391004 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerStarted","Data":"7263cde13925ae3c7d67b8bc6180bda1accc8f2ced8fa93c3aea0d0211094759"} Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.395227 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-459j5" event={"ID":"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92","Type":"ContainerStarted","Data":"4b93a345b80763c54704d57faa06830fee0618b5996fd0dd384828d49fd5a1c3"} Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.398265 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6qj89" event={"ID":"62f6c233-faeb-47e1-8103-4902eaea8d4a","Type":"ContainerStarted","Data":"a45fe580eb50011e8314d6a4e9d7459efe89b3417282abc0f07b1360280af6d2"} Oct 04 04:56:27 crc kubenswrapper[4575]: E1004 04:56:27.401274 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-kg5nx" podUID="ef74f3cc-9958-4de3-a914-12e5453e169d" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.441158 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-647567c5b9-xst9n"] Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.462164 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-647567c5b9-xst9n"] Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.463709 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-459j5" podStartSLOduration=2.570807521 podStartE2EDuration="51.463691683s" podCreationTimestamp="2025-10-04 04:55:36 +0000 UTC" firstStartedPulling="2025-10-04 04:55:37.827129217 +0000 UTC m=+1289.155688031" lastFinishedPulling="2025-10-04 04:56:26.720013379 +0000 UTC m=+1338.048572193" observedRunningTime="2025-10-04 04:56:27.459734179 +0000 UTC m=+1338.788293003" watchObservedRunningTime="2025-10-04 04:56:27.463691683 +0000 UTC m=+1338.792250497" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.544081 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-6qj89" podStartSLOduration=2.53663965 podStartE2EDuration="51.544058666s" podCreationTimestamp="2025-10-04 04:55:36 +0000 UTC" firstStartedPulling="2025-10-04 04:55:37.86509164 +0000 UTC m=+1289.193650454" lastFinishedPulling="2025-10-04 04:56:26.872510656 +0000 UTC m=+1338.201069470" observedRunningTime="2025-10-04 04:56:27.500218229 +0000 UTC m=+1338.828777053" watchObservedRunningTime="2025-10-04 04:56:27.544058666 +0000 UTC m=+1338.872617480" Oct 04 04:56:27 crc kubenswrapper[4575]: W1004 04:56:27.809511 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca934f38_b0b0_4d26_bf2e_63f983c81851.slice/crio-2c2a842918f9ba8b663ceedff13d185852677dcf0e5d245cb3ee3e8630607a66 WatchSource:0}: Error finding container 2c2a842918f9ba8b663ceedff13d185852677dcf0e5d245cb3ee3e8630607a66: Status 404 returned error can't find the container with id 2c2a842918f9ba8b663ceedff13d185852677dcf0e5d245cb3ee3e8630607a66 Oct 04 04:56:27 crc kubenswrapper[4575]: W1004 04:56:27.816451 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f0573f0_1418_43eb_be82_166fb99ed23d.slice/crio-d56a5ca19304d40eb64c7f7812f453589d44f7c48b6fe6fc999bb409e6a348dd WatchSource:0}: Error finding container d56a5ca19304d40eb64c7f7812f453589d44f7c48b6fe6fc999bb409e6a348dd: Status 404 returned error can't find the container with id d56a5ca19304d40eb64c7f7812f453589d44f7c48b6fe6fc999bb409e6a348dd Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.898358 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gtvfg" Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.991697 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-db-sync-config-data\") pod \"775d96cb-5e3b-437f-b61d-7799e86f2f58\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.991867 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hltsv\" (UniqueName: \"kubernetes.io/projected/775d96cb-5e3b-437f-b61d-7799e86f2f58-kube-api-access-hltsv\") pod \"775d96cb-5e3b-437f-b61d-7799e86f2f58\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.991956 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-config-data\") pod \"775d96cb-5e3b-437f-b61d-7799e86f2f58\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.992024 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-combined-ca-bundle\") pod \"775d96cb-5e3b-437f-b61d-7799e86f2f58\" (UID: \"775d96cb-5e3b-437f-b61d-7799e86f2f58\") " Oct 04 04:56:27 crc kubenswrapper[4575]: I1004 04:56:27.996921 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "775d96cb-5e3b-437f-b61d-7799e86f2f58" (UID: "775d96cb-5e3b-437f-b61d-7799e86f2f58"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.006500 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/775d96cb-5e3b-437f-b61d-7799e86f2f58-kube-api-access-hltsv" (OuterVolumeSpecName: "kube-api-access-hltsv") pod "775d96cb-5e3b-437f-b61d-7799e86f2f58" (UID: "775d96cb-5e3b-437f-b61d-7799e86f2f58"). InnerVolumeSpecName "kube-api-access-hltsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.060552 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "775d96cb-5e3b-437f-b61d-7799e86f2f58" (UID: "775d96cb-5e3b-437f-b61d-7799e86f2f58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.061123 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-config-data" (OuterVolumeSpecName: "config-data") pod "775d96cb-5e3b-437f-b61d-7799e86f2f58" (UID: "775d96cb-5e3b-437f-b61d-7799e86f2f58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.094517 4575 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.094553 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hltsv\" (UniqueName: \"kubernetes.io/projected/775d96cb-5e3b-437f-b61d-7799e86f2f58-kube-api-access-hltsv\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.094568 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.094580 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775d96cb-5e3b-437f-b61d-7799e86f2f58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.419638 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-99bks" event={"ID":"ca934f38-b0b0-4d26-bf2e-63f983c81851","Type":"ContainerStarted","Data":"2dab7ad46d64d8b960c4f7b4044c68c2595dbab9cc061d49f6dd8f075dc41c0c"} Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.419974 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-99bks" event={"ID":"ca934f38-b0b0-4d26-bf2e-63f983c81851","Type":"ContainerStarted","Data":"2c2a842918f9ba8b663ceedff13d185852677dcf0e5d245cb3ee3e8630607a66"} Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.434067 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89017745-df9d-4bae-ad06-786a6b653ce8","Type":"ContainerStarted","Data":"c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0"} Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.452541 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gtvfg" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.452710 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gtvfg" event={"ID":"775d96cb-5e3b-437f-b61d-7799e86f2f58","Type":"ContainerDied","Data":"dfcbe4cddb62c444eef9edf121c5a3903a63dd2d748a3393dd625afbcbee76fb"} Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.452762 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfcbe4cddb62c444eef9edf121c5a3903a63dd2d748a3393dd625afbcbee76fb" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.481954 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerStarted","Data":"fa879635a88d7991c3adb1c644cf46f55b5522a646d4725f1a946e36c99c339d"} Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.482007 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerStarted","Data":"d943ddf17584717bada181edfe28e9bf789e3b389b2a38e1aa4033b28572a55c"} Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.482022 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerStarted","Data":"d56a5ca19304d40eb64c7f7812f453589d44f7c48b6fe6fc999bb409e6a348dd"} Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.492564 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerStarted","Data":"d57267105118ef90e3d164b8092e906f1721d7d460bbbe2d30820ae6ca7a010d"} Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.492618 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerStarted","Data":"bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149"} Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.526754 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-644bf5cdd4-pwdw9" podStartSLOduration=43.526730258 podStartE2EDuration="43.526730258s" podCreationTimestamp="2025-10-04 04:55:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:28.51919528 +0000 UTC m=+1339.847754094" watchObservedRunningTime="2025-10-04 04:56:28.526730258 +0000 UTC m=+1339.855289072" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.529201 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-99bks" podStartSLOduration=35.529187539 podStartE2EDuration="35.529187539s" podCreationTimestamp="2025-10-04 04:55:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:28.452511202 +0000 UTC m=+1339.781070016" watchObservedRunningTime="2025-10-04 04:56:28.529187539 +0000 UTC m=+1339.857746353" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.557455 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-b86b879b4-tlf42" podStartSLOduration=42.467572075 podStartE2EDuration="43.557437215s" podCreationTimestamp="2025-10-04 04:55:45 +0000 UTC" firstStartedPulling="2025-10-04 04:56:26.723225531 +0000 UTC m=+1338.051784345" lastFinishedPulling="2025-10-04 04:56:27.813090671 +0000 UTC m=+1339.141649485" observedRunningTime="2025-10-04 04:56:28.555720946 +0000 UTC m=+1339.884279780" watchObservedRunningTime="2025-10-04 04:56:28.557437215 +0000 UTC m=+1339.885996019" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.905816 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-c926l"] Oct 04 04:56:28 crc kubenswrapper[4575]: E1004 04:56:28.907763 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="init" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.907916 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="init" Oct 04 04:56:28 crc kubenswrapper[4575]: E1004 04:56:28.908009 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="dnsmasq-dns" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.908076 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="dnsmasq-dns" Oct 04 04:56:28 crc kubenswrapper[4575]: E1004 04:56:28.908187 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775d96cb-5e3b-437f-b61d-7799e86f2f58" containerName="glance-db-sync" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.908315 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="775d96cb-5e3b-437f-b61d-7799e86f2f58" containerName="glance-db-sync" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.908672 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dca86a7-57a1-457e-adf9-58882540f233" containerName="dnsmasq-dns" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.908768 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="775d96cb-5e3b-437f-b61d-7799e86f2f58" containerName="glance-db-sync" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.909995 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:28 crc kubenswrapper[4575]: I1004 04:56:28.930074 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-c926l"] Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.009633 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcccb\" (UniqueName: \"kubernetes.io/projected/d1899e27-5c73-4db1-a439-46783fad130d-kube-api-access-tcccb\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.009712 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.009770 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.009843 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.009895 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.009922 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-config\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.111960 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.112166 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.112222 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.112281 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-config\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.112334 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcccb\" (UniqueName: \"kubernetes.io/projected/d1899e27-5c73-4db1-a439-46783fad130d-kube-api-access-tcccb\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.112374 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.113192 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.113192 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.113284 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.113444 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-config\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.113819 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.152132 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcccb\" (UniqueName: \"kubernetes.io/projected/d1899e27-5c73-4db1-a439-46783fad130d-kube-api-access-tcccb\") pod \"dnsmasq-dns-57c957c4ff-c926l\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.231398 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.354189 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9" path="/var/lib/kubelet/pods/1a37ed1f-1285-4ea0-827c-cd6f5ec7a1c9/volumes" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.864030 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.872995 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.878537 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.878962 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.879238 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xtpfm" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.915149 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.932707 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-config-data\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.932804 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.932828 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-scripts\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.932855 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.932885 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8gxd\" (UniqueName: \"kubernetes.io/projected/4b37a6cf-2a40-4063-a359-ac6638357095-kube-api-access-s8gxd\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.932986 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-logs\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.933009 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:29 crc kubenswrapper[4575]: I1004 04:56:29.964073 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-c926l"] Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.034256 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-logs\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.034726 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.034876 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-config-data\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.035074 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.034791 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-logs\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.035086 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.035466 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-scripts\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.035574 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.035700 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8gxd\" (UniqueName: \"kubernetes.io/projected/4b37a6cf-2a40-4063-a359-ac6638357095-kube-api-access-s8gxd\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.038197 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.060215 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.061219 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-scripts\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.061364 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-config-data\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.072396 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8gxd\" (UniqueName: \"kubernetes.io/projected/4b37a6cf-2a40-4063-a359-ac6638357095-kube-api-access-s8gxd\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.087709 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.091037 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.095245 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.099275 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.137977 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.138090 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.138148 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzmh8\" (UniqueName: \"kubernetes.io/projected/f56e68db-7c69-401e-85d2-aed83d7c7a51-kube-api-access-vzmh8\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.138188 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.138255 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-logs\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.138311 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.138334 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.149310 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.205942 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.239796 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.239887 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-logs\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.239945 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.239970 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.240012 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.240075 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.240124 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzmh8\" (UniqueName: \"kubernetes.io/projected/f56e68db-7c69-401e-85d2-aed83d7c7a51-kube-api-access-vzmh8\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.241034 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.241360 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-logs\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.241579 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.251785 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.253806 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.255833 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.273530 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzmh8\" (UniqueName: \"kubernetes.io/projected/f56e68db-7c69-401e-85d2-aed83d7c7a51-kube-api-access-vzmh8\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.326261 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.474579 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.556888 4575 generic.go:334] "Generic (PLEG): container finished" podID="d1899e27-5c73-4db1-a439-46783fad130d" containerID="fdbddc0ae96c0f36ff038f1a964958d7877bfa02acf31aa8209a5b87ea41ce06" exitCode=0 Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.557207 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" event={"ID":"d1899e27-5c73-4db1-a439-46783fad130d","Type":"ContainerDied","Data":"fdbddc0ae96c0f36ff038f1a964958d7877bfa02acf31aa8209a5b87ea41ce06"} Oct 04 04:56:30 crc kubenswrapper[4575]: I1004 04:56:30.557241 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" event={"ID":"d1899e27-5c73-4db1-a439-46783fad130d","Type":"ContainerStarted","Data":"4c394639ec972ee26cd7af7ed411e04fe09b6b2e87131c495d97d23f46045060"} Oct 04 04:56:31 crc kubenswrapper[4575]: I1004 04:56:31.068491 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:56:31 crc kubenswrapper[4575]: W1004 04:56:31.084707 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b37a6cf_2a40_4063_a359_ac6638357095.slice/crio-3d02058c8931e42a1c6d3cf5155758cc280139797abe3d13bd1b6cd1405d01a2 WatchSource:0}: Error finding container 3d02058c8931e42a1c6d3cf5155758cc280139797abe3d13bd1b6cd1405d01a2: Status 404 returned error can't find the container with id 3d02058c8931e42a1c6d3cf5155758cc280139797abe3d13bd1b6cd1405d01a2 Oct 04 04:56:31 crc kubenswrapper[4575]: I1004 04:56:31.328808 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:56:31 crc kubenswrapper[4575]: I1004 04:56:31.595860 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b37a6cf-2a40-4063-a359-ac6638357095","Type":"ContainerStarted","Data":"3d02058c8931e42a1c6d3cf5155758cc280139797abe3d13bd1b6cd1405d01a2"} Oct 04 04:56:31 crc kubenswrapper[4575]: I1004 04:56:31.598314 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f56e68db-7c69-401e-85d2-aed83d7c7a51","Type":"ContainerStarted","Data":"1dab3dff68d0b36a145f7bb920f89641873a264684d7cf5aa20615bf1c150b4b"} Oct 04 04:56:31 crc kubenswrapper[4575]: I1004 04:56:31.601471 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" event={"ID":"d1899e27-5c73-4db1-a439-46783fad130d","Type":"ContainerStarted","Data":"0da0a7a32cc3f718d80a4f68b83dde88bf30c987112dc857e9a3e1cd1357799f"} Oct 04 04:56:31 crc kubenswrapper[4575]: I1004 04:56:31.602497 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:31 crc kubenswrapper[4575]: I1004 04:56:31.638762 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" podStartSLOduration=3.638737604 podStartE2EDuration="3.638737604s" podCreationTimestamp="2025-10-04 04:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:31.62893376 +0000 UTC m=+1342.957492604" watchObservedRunningTime="2025-10-04 04:56:31.638737604 +0000 UTC m=+1342.967296418" Oct 04 04:56:32 crc kubenswrapper[4575]: I1004 04:56:32.176991 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:56:32 crc kubenswrapper[4575]: I1004 04:56:32.284785 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:56:32 crc kubenswrapper[4575]: I1004 04:56:32.622415 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b37a6cf-2a40-4063-a359-ac6638357095","Type":"ContainerStarted","Data":"b70b77efb93c26a7588bba586f83237944b22407a7f6746221e8abb0f5b492d6"} Oct 04 04:56:33 crc kubenswrapper[4575]: I1004 04:56:33.639158 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f56e68db-7c69-401e-85d2-aed83d7c7a51","Type":"ContainerStarted","Data":"6b2446cce394e7d0646c73e7d15d88330c59411da265850510900d3fe348d827"} Oct 04 04:56:33 crc kubenswrapper[4575]: I1004 04:56:33.643349 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b37a6cf-2a40-4063-a359-ac6638357095","Type":"ContainerStarted","Data":"6af4cdb5286e2b799c70cf3756a4fef3b7f612b0346904ce1aafe66266e80caa"} Oct 04 04:56:33 crc kubenswrapper[4575]: I1004 04:56:33.643721 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4b37a6cf-2a40-4063-a359-ac6638357095" containerName="glance-log" containerID="cri-o://b70b77efb93c26a7588bba586f83237944b22407a7f6746221e8abb0f5b492d6" gracePeriod=30 Oct 04 04:56:33 crc kubenswrapper[4575]: I1004 04:56:33.644463 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4b37a6cf-2a40-4063-a359-ac6638357095" containerName="glance-httpd" containerID="cri-o://6af4cdb5286e2b799c70cf3756a4fef3b7f612b0346904ce1aafe66266e80caa" gracePeriod=30 Oct 04 04:56:33 crc kubenswrapper[4575]: I1004 04:56:33.701992 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.701970388 podStartE2EDuration="5.701970388s" podCreationTimestamp="2025-10-04 04:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:33.681217188 +0000 UTC m=+1345.009776002" watchObservedRunningTime="2025-10-04 04:56:33.701970388 +0000 UTC m=+1345.030529202" Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.658837 4575 generic.go:334] "Generic (PLEG): container finished" podID="4b37a6cf-2a40-4063-a359-ac6638357095" containerID="6af4cdb5286e2b799c70cf3756a4fef3b7f612b0346904ce1aafe66266e80caa" exitCode=143 Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.659131 4575 generic.go:334] "Generic (PLEG): container finished" podID="4b37a6cf-2a40-4063-a359-ac6638357095" containerID="b70b77efb93c26a7588bba586f83237944b22407a7f6746221e8abb0f5b492d6" exitCode=143 Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.658873 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b37a6cf-2a40-4063-a359-ac6638357095","Type":"ContainerDied","Data":"6af4cdb5286e2b799c70cf3756a4fef3b7f612b0346904ce1aafe66266e80caa"} Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.659193 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b37a6cf-2a40-4063-a359-ac6638357095","Type":"ContainerDied","Data":"b70b77efb93c26a7588bba586f83237944b22407a7f6746221e8abb0f5b492d6"} Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.665911 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f56e68db-7c69-401e-85d2-aed83d7c7a51","Type":"ContainerStarted","Data":"7c12697af7b310bb9306affa7bc9bc348df4cc30f4c35d5fcce18b4c3ec029c3"} Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.665963 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerName="glance-log" containerID="cri-o://6b2446cce394e7d0646c73e7d15d88330c59411da265850510900d3fe348d827" gracePeriod=30 Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.666062 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerName="glance-httpd" containerID="cri-o://7c12697af7b310bb9306affa7bc9bc348df4cc30f4c35d5fcce18b4c3ec029c3" gracePeriod=30 Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.668569 4575 generic.go:334] "Generic (PLEG): container finished" podID="fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" containerID="4b93a345b80763c54704d57faa06830fee0618b5996fd0dd384828d49fd5a1c3" exitCode=0 Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.668621 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-459j5" event={"ID":"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92","Type":"ContainerDied","Data":"4b93a345b80763c54704d57faa06830fee0618b5996fd0dd384828d49fd5a1c3"} Oct 04 04:56:34 crc kubenswrapper[4575]: I1004 04:56:34.693767 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.693748623 podStartE2EDuration="5.693748623s" podCreationTimestamp="2025-10-04 04:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:34.689518941 +0000 UTC m=+1346.018077755" watchObservedRunningTime="2025-10-04 04:56:34.693748623 +0000 UTC m=+1346.022307437" Oct 04 04:56:35 crc kubenswrapper[4575]: I1004 04:56:35.557675 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:56:35 crc kubenswrapper[4575]: I1004 04:56:35.558323 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:56:35 crc kubenswrapper[4575]: I1004 04:56:35.680234 4575 generic.go:334] "Generic (PLEG): container finished" podID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerID="7c12697af7b310bb9306affa7bc9bc348df4cc30f4c35d5fcce18b4c3ec029c3" exitCode=0 Oct 04 04:56:35 crc kubenswrapper[4575]: I1004 04:56:35.680268 4575 generic.go:334] "Generic (PLEG): container finished" podID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerID="6b2446cce394e7d0646c73e7d15d88330c59411da265850510900d3fe348d827" exitCode=143 Oct 04 04:56:35 crc kubenswrapper[4575]: I1004 04:56:35.680314 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f56e68db-7c69-401e-85d2-aed83d7c7a51","Type":"ContainerDied","Data":"7c12697af7b310bb9306affa7bc9bc348df4cc30f4c35d5fcce18b4c3ec029c3"} Oct 04 04:56:35 crc kubenswrapper[4575]: I1004 04:56:35.680354 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f56e68db-7c69-401e-85d2-aed83d7c7a51","Type":"ContainerDied","Data":"6b2446cce394e7d0646c73e7d15d88330c59411da265850510900d3fe348d827"} Oct 04 04:56:35 crc kubenswrapper[4575]: I1004 04:56:35.713603 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:56:35 crc kubenswrapper[4575]: I1004 04:56:35.713666 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:56:38 crc kubenswrapper[4575]: I1004 04:56:38.715479 4575 generic.go:334] "Generic (PLEG): container finished" podID="ca934f38-b0b0-4d26-bf2e-63f983c81851" containerID="2dab7ad46d64d8b960c4f7b4044c68c2595dbab9cc061d49f6dd8f075dc41c0c" exitCode=0 Oct 04 04:56:38 crc kubenswrapper[4575]: I1004 04:56:38.715544 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-99bks" event={"ID":"ca934f38-b0b0-4d26-bf2e-63f983c81851","Type":"ContainerDied","Data":"2dab7ad46d64d8b960c4f7b4044c68c2595dbab9cc061d49f6dd8f075dc41c0c"} Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.141945 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-459j5" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.234863 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.298184 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-logs\") pod \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.298259 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-config-data\") pod \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.298318 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-scripts\") pod \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.298404 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzfdm\" (UniqueName: \"kubernetes.io/projected/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-kube-api-access-hzfdm\") pod \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.298445 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-combined-ca-bundle\") pod \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\" (UID: \"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.342979 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-logs" (OuterVolumeSpecName: "logs") pod "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" (UID: "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.351044 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-scripts" (OuterVolumeSpecName: "scripts") pod "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" (UID: "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.384019 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-kube-api-access-hzfdm" (OuterVolumeSpecName: "kube-api-access-hzfdm") pod "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" (UID: "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92"). InnerVolumeSpecName "kube-api-access-hzfdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.401256 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.401344 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.401361 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzfdm\" (UniqueName: \"kubernetes.io/projected/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-kube-api-access-hzfdm\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.429091 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-cwbzl"] Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.429387 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" podUID="74dc644a-498d-47b5-af2d-164d5eee997a" containerName="dnsmasq-dns" containerID="cri-o://08c4ccb92fab96b41417b0a4ae95f796b798dee566942cf7a1a9aea9ea4c94b5" gracePeriod=10 Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.511775 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-config-data" (OuterVolumeSpecName: "config-data") pod "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" (UID: "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.529938 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" (UID: "fc2d4d08-1560-4aa3-8a6e-5350abdd3b92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.547611 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.648982 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.717067 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.789745 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-459j5" event={"ID":"fc2d4d08-1560-4aa3-8a6e-5350abdd3b92","Type":"ContainerDied","Data":"1dda133b2d3dcc676eac03d1beb1ed0afc02a66e634dc4304c63f748a7ad338c"} Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.789785 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dda133b2d3dcc676eac03d1beb1ed0afc02a66e634dc4304c63f748a7ad338c" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.789890 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-459j5" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.806260 4575 generic.go:334] "Generic (PLEG): container finished" podID="62f6c233-faeb-47e1-8103-4902eaea8d4a" containerID="a45fe580eb50011e8314d6a4e9d7459efe89b3417282abc0f07b1360280af6d2" exitCode=0 Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.806443 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6qj89" event={"ID":"62f6c233-faeb-47e1-8103-4902eaea8d4a","Type":"ContainerDied","Data":"a45fe580eb50011e8314d6a4e9d7459efe89b3417282abc0f07b1360280af6d2"} Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.819854 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.820306 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b37a6cf-2a40-4063-a359-ac6638357095","Type":"ContainerDied","Data":"3d02058c8931e42a1c6d3cf5155758cc280139797abe3d13bd1b6cd1405d01a2"} Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.820341 4575 scope.go:117] "RemoveContainer" containerID="6af4cdb5286e2b799c70cf3756a4fef3b7f612b0346904ce1aafe66266e80caa" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.832616 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89017745-df9d-4bae-ad06-786a6b653ce8","Type":"ContainerStarted","Data":"bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66"} Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.851787 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-config-data\") pod \"4b37a6cf-2a40-4063-a359-ac6638357095\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.852155 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-combined-ca-bundle\") pod \"4b37a6cf-2a40-4063-a359-ac6638357095\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.852205 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-scripts\") pod \"4b37a6cf-2a40-4063-a359-ac6638357095\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.852226 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-logs\") pod \"4b37a6cf-2a40-4063-a359-ac6638357095\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.852297 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8gxd\" (UniqueName: \"kubernetes.io/projected/4b37a6cf-2a40-4063-a359-ac6638357095-kube-api-access-s8gxd\") pod \"4b37a6cf-2a40-4063-a359-ac6638357095\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.852347 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-httpd-run\") pod \"4b37a6cf-2a40-4063-a359-ac6638357095\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.852410 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"4b37a6cf-2a40-4063-a359-ac6638357095\" (UID: \"4b37a6cf-2a40-4063-a359-ac6638357095\") " Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.862249 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-logs" (OuterVolumeSpecName: "logs") pod "4b37a6cf-2a40-4063-a359-ac6638357095" (UID: "4b37a6cf-2a40-4063-a359-ac6638357095"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.862543 4575 generic.go:334] "Generic (PLEG): container finished" podID="74dc644a-498d-47b5-af2d-164d5eee997a" containerID="08c4ccb92fab96b41417b0a4ae95f796b798dee566942cf7a1a9aea9ea4c94b5" exitCode=0 Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.862855 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" event={"ID":"74dc644a-498d-47b5-af2d-164d5eee997a","Type":"ContainerDied","Data":"08c4ccb92fab96b41417b0a4ae95f796b798dee566942cf7a1a9aea9ea4c94b5"} Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.863496 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-scripts" (OuterVolumeSpecName: "scripts") pod "4b37a6cf-2a40-4063-a359-ac6638357095" (UID: "4b37a6cf-2a40-4063-a359-ac6638357095"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.864357 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4b37a6cf-2a40-4063-a359-ac6638357095" (UID: "4b37a6cf-2a40-4063-a359-ac6638357095"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.875037 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "4b37a6cf-2a40-4063-a359-ac6638357095" (UID: "4b37a6cf-2a40-4063-a359-ac6638357095"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.887655 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b37a6cf-2a40-4063-a359-ac6638357095-kube-api-access-s8gxd" (OuterVolumeSpecName: "kube-api-access-s8gxd") pod "4b37a6cf-2a40-4063-a359-ac6638357095" (UID: "4b37a6cf-2a40-4063-a359-ac6638357095"). InnerVolumeSpecName "kube-api-access-s8gxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.928425 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b37a6cf-2a40-4063-a359-ac6638357095" (UID: "4b37a6cf-2a40-4063-a359-ac6638357095"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.938129 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-config-data" (OuterVolumeSpecName: "config-data") pod "4b37a6cf-2a40-4063-a359-ac6638357095" (UID: "4b37a6cf-2a40-4063-a359-ac6638357095"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.958402 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.958446 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.958459 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.958469 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b37a6cf-2a40-4063-a359-ac6638357095-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.958479 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8gxd\" (UniqueName: \"kubernetes.io/projected/4b37a6cf-2a40-4063-a359-ac6638357095-kube-api-access-s8gxd\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.958491 4575 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b37a6cf-2a40-4063-a359-ac6638357095-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:39 crc kubenswrapper[4575]: I1004 04:56:39.958520 4575 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.030012 4575 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.059933 4575 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.111964 4575 scope.go:117] "RemoveContainer" containerID="b70b77efb93c26a7588bba586f83237944b22407a7f6746221e8abb0f5b492d6" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.184880 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.212876 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.266072 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:56:40 crc kubenswrapper[4575]: E1004 04:56:40.266533 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b37a6cf-2a40-4063-a359-ac6638357095" containerName="glance-log" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.266574 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b37a6cf-2a40-4063-a359-ac6638357095" containerName="glance-log" Oct 04 04:56:40 crc kubenswrapper[4575]: E1004 04:56:40.266625 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" containerName="placement-db-sync" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.266635 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" containerName="placement-db-sync" Oct 04 04:56:40 crc kubenswrapper[4575]: E1004 04:56:40.266651 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b37a6cf-2a40-4063-a359-ac6638357095" containerName="glance-httpd" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.266659 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b37a6cf-2a40-4063-a359-ac6638357095" containerName="glance-httpd" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.266865 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" containerName="placement-db-sync" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.266881 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b37a6cf-2a40-4063-a359-ac6638357095" containerName="glance-log" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.266894 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b37a6cf-2a40-4063-a359-ac6638357095" containerName="glance-httpd" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.279659 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.293734 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.299114 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.299385 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.354605 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-745885db44-gjnjj"] Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.358119 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.370248 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.370305 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-7pkll" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.370908 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.371077 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.371299 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.414178 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-745885db44-gjnjj"] Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.472570 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgcwk\" (UniqueName: \"kubernetes.io/projected/86986b78-6e24-4238-a515-9e5176d6847a-kube-api-access-cgcwk\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.472798 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-internal-tls-certs\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.472821 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.472853 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.472882 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-public-tls-certs\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.472912 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlxn9\" (UniqueName: \"kubernetes.io/projected/248dffc1-b899-42d0-8591-9b232d821972-kube-api-access-xlxn9\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.472935 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86986b78-6e24-4238-a515-9e5176d6847a-logs\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.472961 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-logs\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.472987 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.473020 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-combined-ca-bundle\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.473051 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-scripts\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.473103 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-config-data\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.473121 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-scripts\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.473145 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-config-data\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.473188 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.583726 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-public-tls-certs\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.583772 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlxn9\" (UniqueName: \"kubernetes.io/projected/248dffc1-b899-42d0-8591-9b232d821972-kube-api-access-xlxn9\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.583788 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86986b78-6e24-4238-a515-9e5176d6847a-logs\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.583836 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-logs\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.583878 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.583930 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-combined-ca-bundle\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.583956 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-scripts\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.584082 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-config-data\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.584105 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-scripts\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.584138 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-config-data\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.584209 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.584256 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgcwk\" (UniqueName: \"kubernetes.io/projected/86986b78-6e24-4238-a515-9e5176d6847a-kube-api-access-cgcwk\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.584290 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-internal-tls-certs\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.584306 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.584350 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.584828 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.589167 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-logs\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.589821 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86986b78-6e24-4238-a515-9e5176d6847a-logs\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.591041 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.607099 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-combined-ca-bundle\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.614494 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-config-data\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.640502 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-config-data\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.641362 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-public-tls-certs\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.659908 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-scripts\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.660494 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-internal-tls-certs\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.660909 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86986b78-6e24-4238-a515-9e5176d6847a-scripts\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.661080 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.661559 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.666319 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlxn9\" (UniqueName: \"kubernetes.io/projected/248dffc1-b899-42d0-8591-9b232d821972-kube-api-access-xlxn9\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.666526 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgcwk\" (UniqueName: \"kubernetes.io/projected/86986b78-6e24-4238-a515-9e5176d6847a-kube-api-access-cgcwk\") pod \"placement-745885db44-gjnjj\" (UID: \"86986b78-6e24-4238-a515-9e5176d6847a\") " pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.667386 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.683482 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.689301 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.721886 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-99bks" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.727362 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800287 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzmh8\" (UniqueName: \"kubernetes.io/projected/f56e68db-7c69-401e-85d2-aed83d7c7a51-kube-api-access-vzmh8\") pod \"f56e68db-7c69-401e-85d2-aed83d7c7a51\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800336 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-config-data\") pod \"f56e68db-7c69-401e-85d2-aed83d7c7a51\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800408 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"f56e68db-7c69-401e-85d2-aed83d7c7a51\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800435 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-config\") pod \"74dc644a-498d-47b5-af2d-164d5eee997a\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800501 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-swift-storage-0\") pod \"74dc644a-498d-47b5-af2d-164d5eee997a\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800567 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv57g\" (UniqueName: \"kubernetes.io/projected/74dc644a-498d-47b5-af2d-164d5eee997a-kube-api-access-tv57g\") pod \"74dc644a-498d-47b5-af2d-164d5eee997a\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800639 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-svc\") pod \"74dc644a-498d-47b5-af2d-164d5eee997a\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800682 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-logs\") pod \"f56e68db-7c69-401e-85d2-aed83d7c7a51\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800699 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-scripts\") pod \"f56e68db-7c69-401e-85d2-aed83d7c7a51\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800720 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-combined-ca-bundle\") pod \"f56e68db-7c69-401e-85d2-aed83d7c7a51\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800761 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-httpd-run\") pod \"f56e68db-7c69-401e-85d2-aed83d7c7a51\" (UID: \"f56e68db-7c69-401e-85d2-aed83d7c7a51\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800794 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-sb\") pod \"74dc644a-498d-47b5-af2d-164d5eee997a\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.800852 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-nb\") pod \"74dc644a-498d-47b5-af2d-164d5eee997a\" (UID: \"74dc644a-498d-47b5-af2d-164d5eee997a\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.802321 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-logs" (OuterVolumeSpecName: "logs") pod "f56e68db-7c69-401e-85d2-aed83d7c7a51" (UID: "f56e68db-7c69-401e-85d2-aed83d7c7a51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.802713 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f56e68db-7c69-401e-85d2-aed83d7c7a51" (UID: "f56e68db-7c69-401e-85d2-aed83d7c7a51"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.822857 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.822903 4575 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f56e68db-7c69-401e-85d2-aed83d7c7a51-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.826740 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-scripts" (OuterVolumeSpecName: "scripts") pod "f56e68db-7c69-401e-85d2-aed83d7c7a51" (UID: "f56e68db-7c69-401e-85d2-aed83d7c7a51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.826964 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74dc644a-498d-47b5-af2d-164d5eee997a-kube-api-access-tv57g" (OuterVolumeSpecName: "kube-api-access-tv57g") pod "74dc644a-498d-47b5-af2d-164d5eee997a" (UID: "74dc644a-498d-47b5-af2d-164d5eee997a"). InnerVolumeSpecName "kube-api-access-tv57g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.829236 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56e68db-7c69-401e-85d2-aed83d7c7a51-kube-api-access-vzmh8" (OuterVolumeSpecName: "kube-api-access-vzmh8") pod "f56e68db-7c69-401e-85d2-aed83d7c7a51" (UID: "f56e68db-7c69-401e-85d2-aed83d7c7a51"). InnerVolumeSpecName "kube-api-access-vzmh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.842809 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "f56e68db-7c69-401e-85d2-aed83d7c7a51" (UID: "f56e68db-7c69-401e-85d2-aed83d7c7a51"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.910273 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" event={"ID":"74dc644a-498d-47b5-af2d-164d5eee997a","Type":"ContainerDied","Data":"fc6790f29ffd6bbd941e75b6bc321187af2b6ab61ec7dd0ea7f44a1ec5dcc31a"} Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.911414 4575 scope.go:117] "RemoveContainer" containerID="08c4ccb92fab96b41417b0a4ae95f796b798dee566942cf7a1a9aea9ea4c94b5" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.910417 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-cwbzl" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.915927 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-99bks" event={"ID":"ca934f38-b0b0-4d26-bf2e-63f983c81851","Type":"ContainerDied","Data":"2c2a842918f9ba8b663ceedff13d185852677dcf0e5d245cb3ee3e8630607a66"} Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.915963 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c2a842918f9ba8b663ceedff13d185852677dcf0e5d245cb3ee3e8630607a66" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.916049 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-99bks" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.932440 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.933571 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvll6\" (UniqueName: \"kubernetes.io/projected/ca934f38-b0b0-4d26-bf2e-63f983c81851-kube-api-access-dvll6\") pod \"ca934f38-b0b0-4d26-bf2e-63f983c81851\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.933632 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-combined-ca-bundle\") pod \"ca934f38-b0b0-4d26-bf2e-63f983c81851\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.933665 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-credential-keys\") pod \"ca934f38-b0b0-4d26-bf2e-63f983c81851\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.933789 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-config-data\") pod \"ca934f38-b0b0-4d26-bf2e-63f983c81851\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.933953 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-fernet-keys\") pod \"ca934f38-b0b0-4d26-bf2e-63f983c81851\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.933981 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-scripts\") pod \"ca934f38-b0b0-4d26-bf2e-63f983c81851\" (UID: \"ca934f38-b0b0-4d26-bf2e-63f983c81851\") " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.937974 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.938396 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f56e68db-7c69-401e-85d2-aed83d7c7a51","Type":"ContainerDied","Data":"1dab3dff68d0b36a145f7bb920f89641873a264684d7cf5aa20615bf1c150b4b"} Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.946143 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzmh8\" (UniqueName: \"kubernetes.io/projected/f56e68db-7c69-401e-85d2-aed83d7c7a51-kube-api-access-vzmh8\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.946190 4575 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.946203 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv57g\" (UniqueName: \"kubernetes.io/projected/74dc644a-498d-47b5-af2d-164d5eee997a-kube-api-access-tv57g\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.946216 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.976410 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ca934f38-b0b0-4d26-bf2e-63f983c81851" (UID: "ca934f38-b0b0-4d26-bf2e-63f983c81851"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.976582 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca934f38-b0b0-4d26-bf2e-63f983c81851-kube-api-access-dvll6" (OuterVolumeSpecName: "kube-api-access-dvll6") pod "ca934f38-b0b0-4d26-bf2e-63f983c81851" (UID: "ca934f38-b0b0-4d26-bf2e-63f983c81851"). InnerVolumeSpecName "kube-api-access-dvll6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.976892 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f56e68db-7c69-401e-85d2-aed83d7c7a51" (UID: "f56e68db-7c69-401e-85d2-aed83d7c7a51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.990323 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-config-data" (OuterVolumeSpecName: "config-data") pod "f56e68db-7c69-401e-85d2-aed83d7c7a51" (UID: "f56e68db-7c69-401e-85d2-aed83d7c7a51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.992786 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ca934f38-b0b0-4d26-bf2e-63f983c81851" (UID: "ca934f38-b0b0-4d26-bf2e-63f983c81851"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:40 crc kubenswrapper[4575]: I1004 04:56:40.992895 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-scripts" (OuterVolumeSpecName: "scripts") pod "ca934f38-b0b0-4d26-bf2e-63f983c81851" (UID: "ca934f38-b0b0-4d26-bf2e-63f983c81851"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.006999 4575 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.019715 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "74dc644a-498d-47b5-af2d-164d5eee997a" (UID: "74dc644a-498d-47b5-af2d-164d5eee997a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.060132 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.060218 4575 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.060268 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.060281 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.060295 4575 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.060307 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvll6\" (UniqueName: \"kubernetes.io/projected/ca934f38-b0b0-4d26-bf2e-63f983c81851-kube-api-access-dvll6\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.060335 4575 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.060347 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56e68db-7c69-401e-85d2-aed83d7c7a51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.082230 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "74dc644a-498d-47b5-af2d-164d5eee997a" (UID: "74dc644a-498d-47b5-af2d-164d5eee997a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.101352 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-config" (OuterVolumeSpecName: "config") pod "74dc644a-498d-47b5-af2d-164d5eee997a" (UID: "74dc644a-498d-47b5-af2d-164d5eee997a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.125034 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "74dc644a-498d-47b5-af2d-164d5eee997a" (UID: "74dc644a-498d-47b5-af2d-164d5eee997a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.139553 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca934f38-b0b0-4d26-bf2e-63f983c81851" (UID: "ca934f38-b0b0-4d26-bf2e-63f983c81851"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.148579 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-config-data" (OuterVolumeSpecName: "config-data") pod "ca934f38-b0b0-4d26-bf2e-63f983c81851" (UID: "ca934f38-b0b0-4d26-bf2e-63f983c81851"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.151561 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "74dc644a-498d-47b5-af2d-164d5eee997a" (UID: "74dc644a-498d-47b5-af2d-164d5eee997a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.165097 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.165139 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.165153 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.165190 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.165203 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74dc644a-498d-47b5-af2d-164d5eee997a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.165215 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca934f38-b0b0-4d26-bf2e-63f983c81851-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.343200 4575 scope.go:117] "RemoveContainer" containerID="7f114fbc003f7b8d37daff943283cc53ee1d8a75880f39dfd86d976c49165342" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.454385 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b37a6cf-2a40-4063-a359-ac6638357095" path="/var/lib/kubelet/pods/4b37a6cf-2a40-4063-a359-ac6638357095/volumes" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.499097 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.513429 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.525627 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:56:41 crc kubenswrapper[4575]: E1004 04:56:41.525996 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74dc644a-498d-47b5-af2d-164d5eee997a" containerName="dnsmasq-dns" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.526009 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="74dc644a-498d-47b5-af2d-164d5eee997a" containerName="dnsmasq-dns" Oct 04 04:56:41 crc kubenswrapper[4575]: E1004 04:56:41.526025 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerName="glance-log" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.526031 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerName="glance-log" Oct 04 04:56:41 crc kubenswrapper[4575]: E1004 04:56:41.526053 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74dc644a-498d-47b5-af2d-164d5eee997a" containerName="init" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.526059 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="74dc644a-498d-47b5-af2d-164d5eee997a" containerName="init" Oct 04 04:56:41 crc kubenswrapper[4575]: E1004 04:56:41.526076 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca934f38-b0b0-4d26-bf2e-63f983c81851" containerName="keystone-bootstrap" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.526082 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca934f38-b0b0-4d26-bf2e-63f983c81851" containerName="keystone-bootstrap" Oct 04 04:56:41 crc kubenswrapper[4575]: E1004 04:56:41.526093 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerName="glance-httpd" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.526098 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerName="glance-httpd" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.526248 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerName="glance-httpd" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.526266 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56e68db-7c69-401e-85d2-aed83d7c7a51" containerName="glance-log" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.526279 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="74dc644a-498d-47b5-af2d-164d5eee997a" containerName="dnsmasq-dns" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.526290 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca934f38-b0b0-4d26-bf2e-63f983c81851" containerName="keystone-bootstrap" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.527208 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.540781 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.541052 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.576850 4575 scope.go:117] "RemoveContainer" containerID="7c12697af7b310bb9306affa7bc9bc348df4cc30f4c35d5fcce18b4c3ec029c3" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.593805 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.605454 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-cwbzl"] Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.623317 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.634315 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-cwbzl"] Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.635711 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.635942 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.636179 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chfg6\" (UniqueName: \"kubernetes.io/projected/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-kube-api-access-chfg6\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.636373 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.636544 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-logs\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.637700 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.637910 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.644698 4575 scope.go:117] "RemoveContainer" containerID="6b2446cce394e7d0646c73e7d15d88330c59411da265850510900d3fe348d827" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.666491 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6qj89" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.738935 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-db-sync-config-data\") pod \"62f6c233-faeb-47e1-8103-4902eaea8d4a\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.739847 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-combined-ca-bundle\") pod \"62f6c233-faeb-47e1-8103-4902eaea8d4a\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.739917 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfdtp\" (UniqueName: \"kubernetes.io/projected/62f6c233-faeb-47e1-8103-4902eaea8d4a-kube-api-access-hfdtp\") pod \"62f6c233-faeb-47e1-8103-4902eaea8d4a\" (UID: \"62f6c233-faeb-47e1-8103-4902eaea8d4a\") " Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.740374 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.740524 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.740616 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.740708 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.740745 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chfg6\" (UniqueName: \"kubernetes.io/projected/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-kube-api-access-chfg6\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.740821 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.740873 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-logs\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.740917 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.741877 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.745766 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-logs\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.745867 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.746061 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "62f6c233-faeb-47e1-8103-4902eaea8d4a" (UID: "62f6c233-faeb-47e1-8103-4902eaea8d4a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.753879 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.757329 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.764575 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62f6c233-faeb-47e1-8103-4902eaea8d4a-kube-api-access-hfdtp" (OuterVolumeSpecName: "kube-api-access-hfdtp") pod "62f6c233-faeb-47e1-8103-4902eaea8d4a" (UID: "62f6c233-faeb-47e1-8103-4902eaea8d4a"). InnerVolumeSpecName "kube-api-access-hfdtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.765236 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.775469 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chfg6\" (UniqueName: \"kubernetes.io/projected/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-kube-api-access-chfg6\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.776429 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.818299 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-745885db44-gjnjj"] Oct 04 04:56:41 crc kubenswrapper[4575]: W1004 04:56:41.841758 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86986b78_6e24_4238_a515_9e5176d6847a.slice/crio-55f6a0553e922082f9d91c8537c5f92fedb23c7441938d9e9ea31ca9e63a3829 WatchSource:0}: Error finding container 55f6a0553e922082f9d91c8537c5f92fedb23c7441938d9e9ea31ca9e63a3829: Status 404 returned error can't find the container with id 55f6a0553e922082f9d91c8537c5f92fedb23c7441938d9e9ea31ca9e63a3829 Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.842904 4575 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.842936 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfdtp\" (UniqueName: \"kubernetes.io/projected/62f6c233-faeb-47e1-8103-4902eaea8d4a-kube-api-access-hfdtp\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.845140 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.869207 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62f6c233-faeb-47e1-8103-4902eaea8d4a" (UID: "62f6c233-faeb-47e1-8103-4902eaea8d4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.876210 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.945411 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f6c233-faeb-47e1-8103-4902eaea8d4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.960967 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-745885db44-gjnjj" event={"ID":"86986b78-6e24-4238-a515-9e5176d6847a","Type":"ContainerStarted","Data":"55f6a0553e922082f9d91c8537c5f92fedb23c7441938d9e9ea31ca9e63a3829"} Oct 04 04:56:41 crc kubenswrapper[4575]: I1004 04:56:41.994536 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-864464787f-5jtr5"] Oct 04 04:56:42 crc kubenswrapper[4575]: E1004 04:56:42.000038 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62f6c233-faeb-47e1-8103-4902eaea8d4a" containerName="barbican-db-sync" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.000073 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="62f6c233-faeb-47e1-8103-4902eaea8d4a" containerName="barbican-db-sync" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.000421 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="62f6c233-faeb-47e1-8103-4902eaea8d4a" containerName="barbican-db-sync" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.001321 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6qj89" event={"ID":"62f6c233-faeb-47e1-8103-4902eaea8d4a","Type":"ContainerDied","Data":"ae4e82bea55252f6efc08bc517eb234bfb8319ff1e23a1d674a03ac36c9026a8"} Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.001356 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae4e82bea55252f6efc08bc517eb234bfb8319ff1e23a1d674a03ac36c9026a8" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.001459 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6qj89" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.001641 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.026563 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xdgq4" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.026776 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.026909 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.027011 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.027100 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.027060 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.050984 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-credential-keys\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.051678 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-combined-ca-bundle\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.051780 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-internal-tls-certs\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.051868 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chp7d\" (UniqueName: \"kubernetes.io/projected/04b6e250-1594-4e94-8803-0cf4ed78ebb2-kube-api-access-chp7d\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.052033 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-fernet-keys\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.052167 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-scripts\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.052253 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-config-data\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.069704 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-public-tls-certs\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.097009 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-864464787f-5jtr5"] Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.115866 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:56:42 crc kubenswrapper[4575]: W1004 04:56:42.145712 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod248dffc1_b899_42d0_8591_9b232d821972.slice/crio-e30e8d800b1169083994d3220d66e42909f76b725a80516435c4ce3c262b8c0c WatchSource:0}: Error finding container e30e8d800b1169083994d3220d66e42909f76b725a80516435c4ce3c262b8c0c: Status 404 returned error can't find the container with id e30e8d800b1169083994d3220d66e42909f76b725a80516435c4ce3c262b8c0c Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.173446 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-fernet-keys\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.173522 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-scripts\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.173548 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-config-data\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.173619 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-public-tls-certs\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.173661 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-credential-keys\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.173679 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-combined-ca-bundle\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.173705 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-internal-tls-certs\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.173727 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chp7d\" (UniqueName: \"kubernetes.io/projected/04b6e250-1594-4e94-8803-0cf4ed78ebb2-kube-api-access-chp7d\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.186488 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-fernet-keys\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.190635 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-credential-keys\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.190985 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-scripts\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.191375 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-public-tls-certs\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.191881 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-internal-tls-certs\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.195268 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-combined-ca-bundle\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.201556 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04b6e250-1594-4e94-8803-0cf4ed78ebb2-config-data\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.203203 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chp7d\" (UniqueName: \"kubernetes.io/projected/04b6e250-1594-4e94-8803-0cf4ed78ebb2-kube-api-access-chp7d\") pod \"keystone-864464787f-5jtr5\" (UID: \"04b6e250-1594-4e94-8803-0cf4ed78ebb2\") " pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.390270 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6d6b4c688c-g2g6g"] Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.392384 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.426788 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-pmrbn" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.428034 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.459296 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6d6b4c688c-g2g6g"] Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.464106 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.479409 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-config-data-custom\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.479477 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-combined-ca-bundle\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.479509 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-config-data\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.479534 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kln67\" (UniqueName: \"kubernetes.io/projected/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-kube-api-access-kln67\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.479551 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-logs\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.487530 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.581785 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-config-data-custom\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.581868 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-combined-ca-bundle\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.581906 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-config-data\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.581941 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kln67\" (UniqueName: \"kubernetes.io/projected/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-kube-api-access-kln67\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.581970 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-logs\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.582538 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-logs\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.607197 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-config-data-custom\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.608837 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-config-data\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.610153 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-combined-ca-bundle\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.677320 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kln67\" (UniqueName: \"kubernetes.io/projected/7d9d9151-3e7f-473a-8c53-6ca926cadb6f-kube-api-access-kln67\") pod \"barbican-worker-6d6b4c688c-g2g6g\" (UID: \"7d9d9151-3e7f-473a-8c53-6ca926cadb6f\") " pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.718767 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-c74ll"] Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.720679 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.784634 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.784705 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.784751 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.784781 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mq25\" (UniqueName: \"kubernetes.io/projected/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-kube-api-access-2mq25\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.784806 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-config\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.784880 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.792180 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d6b4c688c-g2g6g" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.808846 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6d899d7ff8-lv6wl"] Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.810421 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.833314 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.896754 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-config-data\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.896824 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-config-data-custom\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.896863 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.896906 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.896955 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.896989 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-logs\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.897015 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtt8z\" (UniqueName: \"kubernetes.io/projected/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-kube-api-access-wtt8z\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.897046 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.897083 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mq25\" (UniqueName: \"kubernetes.io/projected/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-kube-api-access-2mq25\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.897124 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-config\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.897164 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-combined-ca-bundle\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.902567 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.910984 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.911779 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-config\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.911817 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-c74ll"] Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.912297 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:42 crc kubenswrapper[4575]: I1004 04:56:42.912477 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.043446 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-logs\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.043554 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtt8z\" (UniqueName: \"kubernetes.io/projected/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-kube-api-access-wtt8z\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.043835 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-combined-ca-bundle\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.043940 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-config-data\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.043982 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-config-data-custom\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.044837 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mq25\" (UniqueName: \"kubernetes.io/projected/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-kube-api-access-2mq25\") pod \"dnsmasq-dns-6d66f584d7-c74ll\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.061873 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-logs\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.061927 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d899d7ff8-lv6wl"] Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.116540 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.167911 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-config-data-custom\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.179823 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtt8z\" (UniqueName: \"kubernetes.io/projected/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-kube-api-access-wtt8z\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.182292 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-config-data\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.204175 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83c5777-2ecd-4b0a-a710-4944a1e33ee4-combined-ca-bundle\") pod \"barbican-keystone-listener-6d899d7ff8-lv6wl\" (UID: \"f83c5777-2ecd-4b0a-a710-4944a1e33ee4\") " pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.221249 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-745885db44-gjnjj" event={"ID":"86986b78-6e24-4238-a515-9e5176d6847a","Type":"ContainerStarted","Data":"b2bfb4946b073116d21f144ce528d69d3b4001c674ba0a03090f2441c5662363"} Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.248074 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.276568 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kg5nx" event={"ID":"ef74f3cc-9958-4de3-a914-12e5453e169d","Type":"ContainerStarted","Data":"dd502cb233d95c2c516c52699cbfd531bb3844e95479f6d08ba7fc1bc7d0794a"} Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.389191 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74dc644a-498d-47b5-af2d-164d5eee997a" path="/var/lib/kubelet/pods/74dc644a-498d-47b5-af2d-164d5eee997a/volumes" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.396353 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56e68db-7c69-401e-85d2-aed83d7c7a51" path="/var/lib/kubelet/pods/f56e68db-7c69-401e-85d2-aed83d7c7a51/volumes" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.399030 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"248dffc1-b899-42d0-8591-9b232d821972","Type":"ContainerStarted","Data":"e30e8d800b1169083994d3220d66e42909f76b725a80516435c4ce3c262b8c0c"} Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.460798 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-kg5nx" podStartSLOduration=5.100389189 podStartE2EDuration="1m7.460737816s" podCreationTimestamp="2025-10-04 04:55:36 +0000 UTC" firstStartedPulling="2025-10-04 04:55:37.615790213 +0000 UTC m=+1288.944349027" lastFinishedPulling="2025-10-04 04:56:39.97613884 +0000 UTC m=+1351.304697654" observedRunningTime="2025-10-04 04:56:43.379676643 +0000 UTC m=+1354.708235477" watchObservedRunningTime="2025-10-04 04:56:43.460737816 +0000 UTC m=+1354.789296630" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.478729 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-765dfcdc6b-ctzxp"] Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.482434 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.514002 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.562040 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-765dfcdc6b-ctzxp"] Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.571634 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.597681 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data-custom\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.601902 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnl6q\" (UniqueName: \"kubernetes.io/projected/98eae969-87b5-45fe-ac11-eecaacb4a459-kube-api-access-rnl6q\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.601964 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-combined-ca-bundle\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.602094 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98eae969-87b5-45fe-ac11-eecaacb4a459-logs\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.602131 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: W1004 04:56:43.631917 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1375b5e3_0250_4fe2_8d8d_e07c4d7afae5.slice/crio-50b85a761779110bffd0df4738de47dc3164a1e72b0f50f113896a5c6bc51f98 WatchSource:0}: Error finding container 50b85a761779110bffd0df4738de47dc3164a1e72b0f50f113896a5c6bc51f98: Status 404 returned error can't find the container with id 50b85a761779110bffd0df4738de47dc3164a1e72b0f50f113896a5c6bc51f98 Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.704423 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98eae969-87b5-45fe-ac11-eecaacb4a459-logs\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.704463 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.704522 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data-custom\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.704557 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnl6q\" (UniqueName: \"kubernetes.io/projected/98eae969-87b5-45fe-ac11-eecaacb4a459-kube-api-access-rnl6q\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.704599 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-combined-ca-bundle\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.705607 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98eae969-87b5-45fe-ac11-eecaacb4a459-logs\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.713980 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-864464787f-5jtr5"] Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.727527 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data-custom\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.734531 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.738770 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-combined-ca-bundle\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.751352 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnl6q\" (UniqueName: \"kubernetes.io/projected/98eae969-87b5-45fe-ac11-eecaacb4a459-kube-api-access-rnl6q\") pod \"barbican-api-765dfcdc6b-ctzxp\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:43 crc kubenswrapper[4575]: W1004 04:56:43.851857 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04b6e250_1594_4e94_8803_0cf4ed78ebb2.slice/crio-bbe7de91b2f4514b2c6fd5618fef054181450d1bdcac66014ca5bd8ef56e9d8c WatchSource:0}: Error finding container bbe7de91b2f4514b2c6fd5618fef054181450d1bdcac66014ca5bd8ef56e9d8c: Status 404 returned error can't find the container with id bbe7de91b2f4514b2c6fd5618fef054181450d1bdcac66014ca5bd8ef56e9d8c Oct 04 04:56:43 crc kubenswrapper[4575]: I1004 04:56:43.886038 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:44 crc kubenswrapper[4575]: I1004 04:56:44.097923 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6d6b4c688c-g2g6g"] Oct 04 04:56:44 crc kubenswrapper[4575]: I1004 04:56:44.378491 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-c74ll"] Oct 04 04:56:44 crc kubenswrapper[4575]: I1004 04:56:44.392843 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d6b4c688c-g2g6g" event={"ID":"7d9d9151-3e7f-473a-8c53-6ca926cadb6f","Type":"ContainerStarted","Data":"66d853eef662126afcf6cb7e477b2c11f606207fbbcdb8eca5ab6085c1cf6bcf"} Oct 04 04:56:44 crc kubenswrapper[4575]: I1004 04:56:44.409877 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5","Type":"ContainerStarted","Data":"50b85a761779110bffd0df4738de47dc3164a1e72b0f50f113896a5c6bc51f98"} Oct 04 04:56:44 crc kubenswrapper[4575]: I1004 04:56:44.415701 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-864464787f-5jtr5" event={"ID":"04b6e250-1594-4e94-8803-0cf4ed78ebb2","Type":"ContainerStarted","Data":"bbe7de91b2f4514b2c6fd5618fef054181450d1bdcac66014ca5bd8ef56e9d8c"} Oct 04 04:56:44 crc kubenswrapper[4575]: I1004 04:56:44.599716 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d899d7ff8-lv6wl"] Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.022203 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-765dfcdc6b-ctzxp"] Oct 04 04:56:45 crc kubenswrapper[4575]: W1004 04:56:45.148783 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98eae969_87b5_45fe_ac11_eecaacb4a459.slice/crio-8b3a70463419547ce8567baed8bb1306a1ff03f0c4a6b98e2e2d3287736f6630 WatchSource:0}: Error finding container 8b3a70463419547ce8567baed8bb1306a1ff03f0c4a6b98e2e2d3287736f6630: Status 404 returned error can't find the container with id 8b3a70463419547ce8567baed8bb1306a1ff03f0c4a6b98e2e2d3287736f6630 Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.449044 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" event={"ID":"1978bad0-ef0f-424e-b8e1-964b99e7b4d6","Type":"ContainerStarted","Data":"8078e3ebc2a3d5d66a3aa221e23d505ab66a75457ae74bbb1b6433aad868d785"} Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.466659 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"248dffc1-b899-42d0-8591-9b232d821972","Type":"ContainerStarted","Data":"7881f5e72ccc072b2c027f6c95b041540b6fda771da724b8c77f85b5f07ebcd5"} Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.491501 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-765dfcdc6b-ctzxp" event={"ID":"98eae969-87b5-45fe-ac11-eecaacb4a459","Type":"ContainerStarted","Data":"8b3a70463419547ce8567baed8bb1306a1ff03f0c4a6b98e2e2d3287736f6630"} Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.499386 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.503423 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" event={"ID":"f83c5777-2ecd-4b0a-a710-4944a1e33ee4","Type":"ContainerStarted","Data":"f44f6e5bd6b675f8e6943a170d222a278ce22e18f4245526cf08feaf5c0583ff"} Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.505883 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-745885db44-gjnjj" event={"ID":"86986b78-6e24-4238-a515-9e5176d6847a","Type":"ContainerStarted","Data":"0cc65b9253c5dbd71b40a2e8a79e6fe6a10dc1de84cecbd193a3193758a8dd4e"} Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.507103 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.507457 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.530890 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-864464787f-5jtr5" podStartSLOduration=4.530866388 podStartE2EDuration="4.530866388s" podCreationTimestamp="2025-10-04 04:56:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:45.524737751 +0000 UTC m=+1356.853296565" watchObservedRunningTime="2025-10-04 04:56:45.530866388 +0000 UTC m=+1356.859425202" Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.568426 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-745885db44-gjnjj" podStartSLOduration=5.568408404 podStartE2EDuration="5.568408404s" podCreationTimestamp="2025-10-04 04:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:45.566349474 +0000 UTC m=+1356.894908288" watchObservedRunningTime="2025-10-04 04:56:45.568408404 +0000 UTC m=+1356.896967218" Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.580376 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:56:45 crc kubenswrapper[4575]: I1004 04:56:45.728699 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.543064 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5","Type":"ContainerStarted","Data":"ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f"} Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.548035 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-765dfcdc6b-ctzxp" event={"ID":"98eae969-87b5-45fe-ac11-eecaacb4a459","Type":"ContainerStarted","Data":"ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269"} Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.561135 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-864464787f-5jtr5" event={"ID":"04b6e250-1594-4e94-8803-0cf4ed78ebb2","Type":"ContainerStarted","Data":"8ab8a5cd4c3ab1be18ac3e525205fd5c62e810e8d620e08b21bcb787c2e45076"} Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.584123 4575 generic.go:334] "Generic (PLEG): container finished" podID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" containerID="622aea6defd2febb4fadcbd6b319150cbc85f8e5a032c33f4536a84809933586" exitCode=0 Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.584250 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" event={"ID":"1978bad0-ef0f-424e-b8e1-964b99e7b4d6","Type":"ContainerDied","Data":"622aea6defd2febb4fadcbd6b319150cbc85f8e5a032c33f4536a84809933586"} Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.759579 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-75d794bf56-cf4hl"] Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.804655 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.806460 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75d794bf56-cf4hl"] Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.812348 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.821120 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.848547 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-config-data\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.848867 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-config-data-custom\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.849021 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-combined-ca-bundle\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.849244 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqqn2\" (UniqueName: \"kubernetes.io/projected/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-kube-api-access-wqqn2\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.849507 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-internal-tls-certs\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.849695 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-public-tls-certs\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.849950 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-logs\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.951744 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-config-data\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.953438 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-config-data-custom\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.953567 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-combined-ca-bundle\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.953692 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqqn2\" (UniqueName: \"kubernetes.io/projected/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-kube-api-access-wqqn2\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.953795 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-public-tls-certs\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.953875 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-internal-tls-certs\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.954049 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-logs\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.954445 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-logs\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.968472 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-config-data\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.969343 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-config-data-custom\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.970334 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-public-tls-certs\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.974272 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-combined-ca-bundle\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.974550 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-internal-tls-certs\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:46 crc kubenswrapper[4575]: I1004 04:56:46.995146 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqqn2\" (UniqueName: \"kubernetes.io/projected/3262a5ef-127a-4d3f-ba16-1287e6e4bd9f-kube-api-access-wqqn2\") pod \"barbican-api-75d794bf56-cf4hl\" (UID: \"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f\") " pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.187714 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.593608 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" event={"ID":"1978bad0-ef0f-424e-b8e1-964b99e7b4d6","Type":"ContainerStarted","Data":"901c7fae38adf9d32365336addacb1a5aa3bba99ee3462617356897f1a80e3d2"} Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.594122 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.597456 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"248dffc1-b899-42d0-8591-9b232d821972","Type":"ContainerStarted","Data":"61770b9da65731707195429d59dbf791815ad68edffb95d4889d2dbc19b49a76"} Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.606149 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5","Type":"ContainerStarted","Data":"72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1"} Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.613674 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-765dfcdc6b-ctzxp" event={"ID":"98eae969-87b5-45fe-ac11-eecaacb4a459","Type":"ContainerStarted","Data":"54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274"} Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.614354 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.614400 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.668885 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.668861884 podStartE2EDuration="7.668861884s" podCreationTimestamp="2025-10-04 04:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:47.660140642 +0000 UTC m=+1358.988699476" watchObservedRunningTime="2025-10-04 04:56:47.668861884 +0000 UTC m=+1358.997420698" Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.671698 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" podStartSLOduration=5.671679425 podStartE2EDuration="5.671679425s" podCreationTimestamp="2025-10-04 04:56:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:47.623373809 +0000 UTC m=+1358.951932633" watchObservedRunningTime="2025-10-04 04:56:47.671679425 +0000 UTC m=+1359.000238259" Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.839994 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-765dfcdc6b-ctzxp" podStartSLOduration=4.83997818 podStartE2EDuration="4.83997818s" podCreationTimestamp="2025-10-04 04:56:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:47.696020419 +0000 UTC m=+1359.024579243" watchObservedRunningTime="2025-10-04 04:56:47.83997818 +0000 UTC m=+1359.168536984" Oct 04 04:56:47 crc kubenswrapper[4575]: I1004 04:56:47.845111 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75d794bf56-cf4hl"] Oct 04 04:56:48 crc kubenswrapper[4575]: I1004 04:56:48.657604 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.657523489 podStartE2EDuration="7.657523489s" podCreationTimestamp="2025-10-04 04:56:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:48.651154545 +0000 UTC m=+1359.979713359" watchObservedRunningTime="2025-10-04 04:56:48.657523489 +0000 UTC m=+1359.986082303" Oct 04 04:56:49 crc kubenswrapper[4575]: I1004 04:56:49.636026 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75d794bf56-cf4hl" event={"ID":"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f","Type":"ContainerStarted","Data":"d197e218c61b42fc55cc57a879d8f4d4600a163653e3f45c52367a58dbed6d5c"} Oct 04 04:56:50 crc kubenswrapper[4575]: I1004 04:56:50.933986 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:56:50 crc kubenswrapper[4575]: I1004 04:56:50.934325 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:56:50 crc kubenswrapper[4575]: I1004 04:56:50.982843 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:56:50 crc kubenswrapper[4575]: I1004 04:56:50.983254 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:56:51 crc kubenswrapper[4575]: I1004 04:56:51.657385 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:56:51 crc kubenswrapper[4575]: I1004 04:56:51.657758 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:56:51 crc kubenswrapper[4575]: I1004 04:56:51.876742 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:51 crc kubenswrapper[4575]: I1004 04:56:51.876802 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:51 crc kubenswrapper[4575]: I1004 04:56:51.921487 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:51 crc kubenswrapper[4575]: I1004 04:56:51.933801 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:52 crc kubenswrapper[4575]: I1004 04:56:52.200750 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:52 crc kubenswrapper[4575]: I1004 04:56:52.667691 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:52 crc kubenswrapper[4575]: I1004 04:56:52.667757 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:53 crc kubenswrapper[4575]: I1004 04:56:53.119788 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:56:53 crc kubenswrapper[4575]: I1004 04:56:53.207862 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-c926l"] Oct 04 04:56:53 crc kubenswrapper[4575]: I1004 04:56:53.208117 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" podUID="d1899e27-5c73-4db1-a439-46783fad130d" containerName="dnsmasq-dns" containerID="cri-o://0da0a7a32cc3f718d80a4f68b83dde88bf30c987112dc857e9a3e1cd1357799f" gracePeriod=10 Oct 04 04:56:53 crc kubenswrapper[4575]: I1004 04:56:53.695743 4575 generic.go:334] "Generic (PLEG): container finished" podID="d1899e27-5c73-4db1-a439-46783fad130d" containerID="0da0a7a32cc3f718d80a4f68b83dde88bf30c987112dc857e9a3e1cd1357799f" exitCode=0 Oct 04 04:56:53 crc kubenswrapper[4575]: I1004 04:56:53.695806 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" event={"ID":"d1899e27-5c73-4db1-a439-46783fad130d","Type":"ContainerDied","Data":"0da0a7a32cc3f718d80a4f68b83dde88bf30c987112dc857e9a3e1cd1357799f"} Oct 04 04:56:53 crc kubenswrapper[4575]: I1004 04:56:53.932811 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-745885db44-gjnjj" Oct 04 04:56:54 crc kubenswrapper[4575]: I1004 04:56:54.237549 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" podUID="d1899e27-5c73-4db1-a439-46783fad130d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: connect: connection refused" Oct 04 04:56:54 crc kubenswrapper[4575]: I1004 04:56:54.710371 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:56:54 crc kubenswrapper[4575]: I1004 04:56:54.716644 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:56:55 crc kubenswrapper[4575]: I1004 04:56:55.559190 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:56:55 crc kubenswrapper[4575]: I1004 04:56:55.714484 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:56:57 crc kubenswrapper[4575]: I1004 04:56:57.970856 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-765dfcdc6b-ctzxp" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:57 crc kubenswrapper[4575]: I1004 04:56:57.970905 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-765dfcdc6b-ctzxp" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:58 crc kubenswrapper[4575]: I1004 04:56:58.402696 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:58 crc kubenswrapper[4575]: I1004 04:56:58.845627 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:56:58 crc kubenswrapper[4575]: I1004 04:56:58.934153 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:56:58 crc kubenswrapper[4575]: I1004 04:56:58.934288 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:56:58 crc kubenswrapper[4575]: I1004 04:56:58.952982 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:58 crc kubenswrapper[4575]: I1004 04:56:58.953092 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:56:58 crc kubenswrapper[4575]: I1004 04:56:58.970778 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:56:59 crc kubenswrapper[4575]: I1004 04:56:59.440082 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:57:01 crc kubenswrapper[4575]: E1004 04:57:01.441434 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Oct 04 04:57:01 crc kubenswrapper[4575]: E1004 04:57:01.441955 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wpdhr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(89017745-df9d-4bae-ad06-786a6b653ce8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:57:01 crc kubenswrapper[4575]: E1004 04:57:01.444963 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.767870 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.818547 4575 generic.go:334] "Generic (PLEG): container finished" podID="ef74f3cc-9958-4de3-a914-12e5453e169d" containerID="dd502cb233d95c2c516c52699cbfd531bb3844e95479f6d08ba7fc1bc7d0794a" exitCode=0 Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.818762 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kg5nx" event={"ID":"ef74f3cc-9958-4de3-a914-12e5453e169d","Type":"ContainerDied","Data":"dd502cb233d95c2c516c52699cbfd531bb3844e95479f6d08ba7fc1bc7d0794a"} Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.844580 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" containerName="ceilometer-notification-agent" containerID="cri-o://c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0" gracePeriod=30 Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.845490 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.848660 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" event={"ID":"d1899e27-5c73-4db1-a439-46783fad130d","Type":"ContainerDied","Data":"4c394639ec972ee26cd7af7ed411e04fe09b6b2e87131c495d97d23f46045060"} Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.848750 4575 scope.go:117] "RemoveContainer" containerID="0da0a7a32cc3f718d80a4f68b83dde88bf30c987112dc857e9a3e1cd1357799f" Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.849002 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" containerName="sg-core" containerID="cri-o://bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66" gracePeriod=30 Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.938041 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-svc\") pod \"d1899e27-5c73-4db1-a439-46783fad130d\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.938127 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-nb\") pod \"d1899e27-5c73-4db1-a439-46783fad130d\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.938172 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-swift-storage-0\") pod \"d1899e27-5c73-4db1-a439-46783fad130d\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.938223 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-sb\") pod \"d1899e27-5c73-4db1-a439-46783fad130d\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.938305 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-config\") pod \"d1899e27-5c73-4db1-a439-46783fad130d\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.938387 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcccb\" (UniqueName: \"kubernetes.io/projected/d1899e27-5c73-4db1-a439-46783fad130d-kube-api-access-tcccb\") pod \"d1899e27-5c73-4db1-a439-46783fad130d\" (UID: \"d1899e27-5c73-4db1-a439-46783fad130d\") " Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.957825 4575 scope.go:117] "RemoveContainer" containerID="fdbddc0ae96c0f36ff038f1a964958d7877bfa02acf31aa8209a5b87ea41ce06" Oct 04 04:57:01 crc kubenswrapper[4575]: I1004 04:57:01.991830 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1899e27-5c73-4db1-a439-46783fad130d-kube-api-access-tcccb" (OuterVolumeSpecName: "kube-api-access-tcccb") pod "d1899e27-5c73-4db1-a439-46783fad130d" (UID: "d1899e27-5c73-4db1-a439-46783fad130d"). InnerVolumeSpecName "kube-api-access-tcccb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.047384 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcccb\" (UniqueName: \"kubernetes.io/projected/d1899e27-5c73-4db1-a439-46783fad130d-kube-api-access-tcccb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.207939 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-config" (OuterVolumeSpecName: "config") pod "d1899e27-5c73-4db1-a439-46783fad130d" (UID: "d1899e27-5c73-4db1-a439-46783fad130d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.234459 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d1899e27-5c73-4db1-a439-46783fad130d" (UID: "d1899e27-5c73-4db1-a439-46783fad130d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.260167 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.260202 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.275174 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d1899e27-5c73-4db1-a439-46783fad130d" (UID: "d1899e27-5c73-4db1-a439-46783fad130d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.302235 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d1899e27-5c73-4db1-a439-46783fad130d" (UID: "d1899e27-5c73-4db1-a439-46783fad130d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.317658 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d1899e27-5c73-4db1-a439-46783fad130d" (UID: "d1899e27-5c73-4db1-a439-46783fad130d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.362391 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.362433 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.362450 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1899e27-5c73-4db1-a439-46783fad130d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.589802 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-c926l"] Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.598307 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-c926l"] Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.856313 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" event={"ID":"f83c5777-2ecd-4b0a-a710-4944a1e33ee4","Type":"ContainerStarted","Data":"88c2a095eea05f97db35a82b446f3c82d0efebfa7bb658b8860ce5bfede739d9"} Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.856715 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" event={"ID":"f83c5777-2ecd-4b0a-a710-4944a1e33ee4","Type":"ContainerStarted","Data":"81f9bef7d54e58ba6e240c65d8631e867db84a7d4f65fbab9342f9ea3f503ac9"} Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.862156 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d6b4c688c-g2g6g" event={"ID":"7d9d9151-3e7f-473a-8c53-6ca926cadb6f","Type":"ContainerStarted","Data":"58b561412601d0c8b67bbc2f2dadc11620b4fda916c8c48ec36aef51247404a8"} Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.862204 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d6b4c688c-g2g6g" event={"ID":"7d9d9151-3e7f-473a-8c53-6ca926cadb6f","Type":"ContainerStarted","Data":"39612de2474618dbde26030ead872254c57ab5eb24c9b7d782f95cbe8629fcf9"} Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.865615 4575 generic.go:334] "Generic (PLEG): container finished" podID="89017745-df9d-4bae-ad06-786a6b653ce8" containerID="bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66" exitCode=2 Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.865709 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89017745-df9d-4bae-ad06-786a6b653ce8","Type":"ContainerDied","Data":"bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66"} Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.868761 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75d794bf56-cf4hl" event={"ID":"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f","Type":"ContainerStarted","Data":"49943b18149c51c5040c6e4b4404e744feac872f354d7fbcc8b0fe151ba05997"} Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.868808 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.868822 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75d794bf56-cf4hl" event={"ID":"3262a5ef-127a-4d3f-ba16-1287e6e4bd9f","Type":"ContainerStarted","Data":"9dbe3cb878548e56ddce4e1d36a1946662d759349c867f068b942bfab46ce521"} Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.868848 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.895011 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6d899d7ff8-lv6wl" podStartSLOduration=4.247028779 podStartE2EDuration="20.894988845s" podCreationTimestamp="2025-10-04 04:56:42 +0000 UTC" firstStartedPulling="2025-10-04 04:56:44.747306742 +0000 UTC m=+1356.075865556" lastFinishedPulling="2025-10-04 04:57:01.395266808 +0000 UTC m=+1372.723825622" observedRunningTime="2025-10-04 04:57:02.879814626 +0000 UTC m=+1374.208373450" watchObservedRunningTime="2025-10-04 04:57:02.894988845 +0000 UTC m=+1374.223547659" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.932980 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-75d794bf56-cf4hl" podStartSLOduration=16.932958642 podStartE2EDuration="16.932958642s" podCreationTimestamp="2025-10-04 04:56:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:02.927600237 +0000 UTC m=+1374.256159061" watchObservedRunningTime="2025-10-04 04:57:02.932958642 +0000 UTC m=+1374.261517456" Oct 04 04:57:02 crc kubenswrapper[4575]: I1004 04:57:02.989184 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6d6b4c688c-g2g6g" podStartSLOduration=3.856079149 podStartE2EDuration="20.989159026s" podCreationTimestamp="2025-10-04 04:56:42 +0000 UTC" firstStartedPulling="2025-10-04 04:56:44.258787142 +0000 UTC m=+1355.587345956" lastFinishedPulling="2025-10-04 04:57:01.391867019 +0000 UTC m=+1372.720425833" observedRunningTime="2025-10-04 04:57:02.985018127 +0000 UTC m=+1374.313576961" watchObservedRunningTime="2025-10-04 04:57:02.989159026 +0000 UTC m=+1374.317717840" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.342740 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1899e27-5c73-4db1-a439-46783fad130d" path="/var/lib/kubelet/pods/d1899e27-5c73-4db1-a439-46783fad130d/volumes" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.501943 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.690822 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2tjm\" (UniqueName: \"kubernetes.io/projected/ef74f3cc-9958-4de3-a914-12e5453e169d-kube-api-access-s2tjm\") pod \"ef74f3cc-9958-4de3-a914-12e5453e169d\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.690914 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-db-sync-config-data\") pod \"ef74f3cc-9958-4de3-a914-12e5453e169d\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.691049 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-combined-ca-bundle\") pod \"ef74f3cc-9958-4de3-a914-12e5453e169d\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.691895 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef74f3cc-9958-4de3-a914-12e5453e169d-etc-machine-id\") pod \"ef74f3cc-9958-4de3-a914-12e5453e169d\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.691970 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef74f3cc-9958-4de3-a914-12e5453e169d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ef74f3cc-9958-4de3-a914-12e5453e169d" (UID: "ef74f3cc-9958-4de3-a914-12e5453e169d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.692051 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-config-data\") pod \"ef74f3cc-9958-4de3-a914-12e5453e169d\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.692411 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-scripts\") pod \"ef74f3cc-9958-4de3-a914-12e5453e169d\" (UID: \"ef74f3cc-9958-4de3-a914-12e5453e169d\") " Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.693383 4575 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef74f3cc-9958-4de3-a914-12e5453e169d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.696852 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ef74f3cc-9958-4de3-a914-12e5453e169d" (UID: "ef74f3cc-9958-4de3-a914-12e5453e169d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.703142 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-scripts" (OuterVolumeSpecName: "scripts") pod "ef74f3cc-9958-4de3-a914-12e5453e169d" (UID: "ef74f3cc-9958-4de3-a914-12e5453e169d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.734784 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef74f3cc-9958-4de3-a914-12e5453e169d-kube-api-access-s2tjm" (OuterVolumeSpecName: "kube-api-access-s2tjm") pod "ef74f3cc-9958-4de3-a914-12e5453e169d" (UID: "ef74f3cc-9958-4de3-a914-12e5453e169d"). InnerVolumeSpecName "kube-api-access-s2tjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.745918 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef74f3cc-9958-4de3-a914-12e5453e169d" (UID: "ef74f3cc-9958-4de3-a914-12e5453e169d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.772180 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-config-data" (OuterVolumeSpecName: "config-data") pod "ef74f3cc-9958-4de3-a914-12e5453e169d" (UID: "ef74f3cc-9958-4de3-a914-12e5453e169d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.795692 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.795731 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2tjm\" (UniqueName: \"kubernetes.io/projected/ef74f3cc-9958-4de3-a914-12e5453e169d-kube-api-access-s2tjm\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.795745 4575 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.795758 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.795770 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef74f3cc-9958-4de3-a914-12e5453e169d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.879671 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kg5nx" Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.882705 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kg5nx" event={"ID":"ef74f3cc-9958-4de3-a914-12e5453e169d","Type":"ContainerDied","Data":"f82c00dce2184a88862328a6fb828aeaa8930d1322e65c3b5e293110beb79f3f"} Oct 04 04:57:03 crc kubenswrapper[4575]: I1004 04:57:03.882755 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f82c00dce2184a88862328a6fb828aeaa8930d1322e65c3b5e293110beb79f3f" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.185870 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:57:04 crc kubenswrapper[4575]: E1004 04:57:04.186287 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1899e27-5c73-4db1-a439-46783fad130d" containerName="dnsmasq-dns" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.186312 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1899e27-5c73-4db1-a439-46783fad130d" containerName="dnsmasq-dns" Oct 04 04:57:04 crc kubenswrapper[4575]: E1004 04:57:04.186330 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1899e27-5c73-4db1-a439-46783fad130d" containerName="init" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.186337 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1899e27-5c73-4db1-a439-46783fad130d" containerName="init" Oct 04 04:57:04 crc kubenswrapper[4575]: E1004 04:57:04.186393 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef74f3cc-9958-4de3-a914-12e5453e169d" containerName="cinder-db-sync" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.186402 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef74f3cc-9958-4de3-a914-12e5453e169d" containerName="cinder-db-sync" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.186634 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1899e27-5c73-4db1-a439-46783fad130d" containerName="dnsmasq-dns" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.186653 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef74f3cc-9958-4de3-a914-12e5453e169d" containerName="cinder-db-sync" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.187817 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.217795 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.217946 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-c79lb" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.218134 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.218253 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.218431 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.239352 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57c957c4ff-c926l" podUID="d1899e27-5c73-4db1-a439-46783fad130d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: i/o timeout" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.323137 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1075fa9-8a84-493b-85f1-2772ebbc4df2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.323188 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-scripts\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.323256 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.323274 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.323318 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.323356 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltwpk\" (UniqueName: \"kubernetes.io/projected/d1075fa9-8a84-493b-85f1-2772ebbc4df2-kube-api-access-ltwpk\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.352408 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-kcbkv"] Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.356647 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.385116 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-kcbkv"] Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.429722 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-config\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.429878 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dbxh\" (UniqueName: \"kubernetes.io/projected/60168709-79af-4397-b804-fbe06d4bc6e3-kube-api-access-5dbxh\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.429937 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-svc\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.429959 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-sb\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.430119 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.431280 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.431642 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.431712 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-swift-storage-0\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.431840 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltwpk\" (UniqueName: \"kubernetes.io/projected/d1075fa9-8a84-493b-85f1-2772ebbc4df2-kube-api-access-ltwpk\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.431888 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-nb\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.432030 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1075fa9-8a84-493b-85f1-2772ebbc4df2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.432090 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-scripts\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.433436 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1075fa9-8a84-493b-85f1-2772ebbc4df2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.445338 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.445816 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.445837 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-scripts\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.457721 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.490182 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltwpk\" (UniqueName: \"kubernetes.io/projected/d1075fa9-8a84-493b-85f1-2772ebbc4df2-kube-api-access-ltwpk\") pod \"cinder-scheduler-0\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.536632 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-config\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.536701 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dbxh\" (UniqueName: \"kubernetes.io/projected/60168709-79af-4397-b804-fbe06d4bc6e3-kube-api-access-5dbxh\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.536722 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-svc\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.536741 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-sb\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.536804 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-swift-storage-0\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.536839 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-nb\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.537619 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-nb\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.538266 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-config\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.539100 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-svc\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.539696 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-sb\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.540122 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-swift-storage-0\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.560025 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.561448 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.564775 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.573222 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.639418 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stlt6\" (UniqueName: \"kubernetes.io/projected/61d80f56-f92c-4725-8c53-71ad25eee1a5-kube-api-access-stlt6\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.639521 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.639636 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61d80f56-f92c-4725-8c53-71ad25eee1a5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.639674 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.639754 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data-custom\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.639796 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-scripts\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.639815 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d80f56-f92c-4725-8c53-71ad25eee1a5-logs\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.650362 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dbxh\" (UniqueName: \"kubernetes.io/projected/60168709-79af-4397-b804-fbe06d4bc6e3-kube-api-access-5dbxh\") pod \"dnsmasq-dns-674b76c99f-kcbkv\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.699059 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.699607 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.741389 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61d80f56-f92c-4725-8c53-71ad25eee1a5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.741451 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.741489 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data-custom\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.741536 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-scripts\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.741558 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d80f56-f92c-4725-8c53-71ad25eee1a5-logs\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.741612 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stlt6\" (UniqueName: \"kubernetes.io/projected/61d80f56-f92c-4725-8c53-71ad25eee1a5-kube-api-access-stlt6\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.741671 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.742848 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d80f56-f92c-4725-8c53-71ad25eee1a5-logs\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.743768 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61d80f56-f92c-4725-8c53-71ad25eee1a5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.749341 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.757764 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data-custom\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.775766 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.776150 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-scripts\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.793870 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stlt6\" (UniqueName: \"kubernetes.io/projected/61d80f56-f92c-4725-8c53-71ad25eee1a5-kube-api-access-stlt6\") pod \"cinder-api-0\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.926393 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.946714 4575 generic.go:334] "Generic (PLEG): container finished" podID="d089c375-6e05-4d74-88e1-d00c8cbddb1f" containerID="3bd213d96e5e86bf506ad4db79bad1f7fde3805d418daa4d347cfcd31d0ed19e" exitCode=0 Oct 04 04:57:04 crc kubenswrapper[4575]: I1004 04:57:04.947323 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-7kr5n" event={"ID":"d089c375-6e05-4d74-88e1-d00c8cbddb1f","Type":"ContainerDied","Data":"3bd213d96e5e86bf506ad4db79bad1f7fde3805d418daa4d347cfcd31d0ed19e"} Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.194186 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.392372 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-kcbkv"] Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.558310 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.558410 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.559289 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"d57267105118ef90e3d164b8092e906f1721d7d460bbbe2d30820ae6ca7a010d"} pod="openstack/horizon-b86b879b4-tlf42" containerMessage="Container horizon failed startup probe, will be restarted" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.559325 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" containerID="cri-o://d57267105118ef90e3d164b8092e906f1721d7d460bbbe2d30820ae6ca7a010d" gracePeriod=30 Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.700527 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.713910 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.713985 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.714742 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"fa879635a88d7991c3adb1c644cf46f55b5522a646d4725f1a946e36c99c339d"} pod="openstack/horizon-644bf5cdd4-pwdw9" containerMessage="Container horizon failed startup probe, will be restarted" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.714780 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" containerID="cri-o://fa879635a88d7991c3adb1c644cf46f55b5522a646d4725f1a946e36c99c339d" gracePeriod=30 Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.729630 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.774350 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-combined-ca-bundle\") pod \"89017745-df9d-4bae-ad06-786a6b653ce8\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.774564 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-scripts\") pod \"89017745-df9d-4bae-ad06-786a6b653ce8\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.774709 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-sg-core-conf-yaml\") pod \"89017745-df9d-4bae-ad06-786a6b653ce8\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.791892 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-scripts" (OuterVolumeSpecName: "scripts") pod "89017745-df9d-4bae-ad06-786a6b653ce8" (UID: "89017745-df9d-4bae-ad06-786a6b653ce8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.813669 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89017745-df9d-4bae-ad06-786a6b653ce8" (UID: "89017745-df9d-4bae-ad06-786a6b653ce8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.833979 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "89017745-df9d-4bae-ad06-786a6b653ce8" (UID: "89017745-df9d-4bae-ad06-786a6b653ce8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.876400 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-run-httpd\") pod \"89017745-df9d-4bae-ad06-786a6b653ce8\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.876459 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-log-httpd\") pod \"89017745-df9d-4bae-ad06-786a6b653ce8\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.876488 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpdhr\" (UniqueName: \"kubernetes.io/projected/89017745-df9d-4bae-ad06-786a6b653ce8-kube-api-access-wpdhr\") pod \"89017745-df9d-4bae-ad06-786a6b653ce8\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.876514 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-config-data\") pod \"89017745-df9d-4bae-ad06-786a6b653ce8\" (UID: \"89017745-df9d-4bae-ad06-786a6b653ce8\") " Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.876931 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.876948 4575 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.876958 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.877568 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "89017745-df9d-4bae-ad06-786a6b653ce8" (UID: "89017745-df9d-4bae-ad06-786a6b653ce8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.877615 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "89017745-df9d-4bae-ad06-786a6b653ce8" (UID: "89017745-df9d-4bae-ad06-786a6b653ce8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.889890 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89017745-df9d-4bae-ad06-786a6b653ce8-kube-api-access-wpdhr" (OuterVolumeSpecName: "kube-api-access-wpdhr") pod "89017745-df9d-4bae-ad06-786a6b653ce8" (UID: "89017745-df9d-4bae-ad06-786a6b653ce8"). InnerVolumeSpecName "kube-api-access-wpdhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.912963 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-config-data" (OuterVolumeSpecName: "config-data") pod "89017745-df9d-4bae-ad06-786a6b653ce8" (UID: "89017745-df9d-4bae-ad06-786a6b653ce8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.983910 4575 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.984149 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpdhr\" (UniqueName: \"kubernetes.io/projected/89017745-df9d-4bae-ad06-786a6b653ce8-kube-api-access-wpdhr\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.984209 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89017745-df9d-4bae-ad06-786a6b653ce8-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.984264 4575 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89017745-df9d-4bae-ad06-786a6b653ce8-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:05 crc kubenswrapper[4575]: I1004 04:57:05.997112 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d1075fa9-8a84-493b-85f1-2772ebbc4df2","Type":"ContainerStarted","Data":"d64d8211704416b22b33137c519222475759a600bdc9f9f44ecf0604bcd173a6"} Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.003090 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"61d80f56-f92c-4725-8c53-71ad25eee1a5","Type":"ContainerStarted","Data":"c4961240e0d2d3b9473edf1457f6ad4206062be0ebad6d9bdadababdaaf03d18"} Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.007994 4575 generic.go:334] "Generic (PLEG): container finished" podID="60168709-79af-4397-b804-fbe06d4bc6e3" containerID="b47b33215697e5938212f2533d8b4e6c47cacdff986366950480273dcca9263d" exitCode=0 Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.008067 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" event={"ID":"60168709-79af-4397-b804-fbe06d4bc6e3","Type":"ContainerDied","Data":"b47b33215697e5938212f2533d8b4e6c47cacdff986366950480273dcca9263d"} Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.008101 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" event={"ID":"60168709-79af-4397-b804-fbe06d4bc6e3","Type":"ContainerStarted","Data":"e0f3c5409003507da7e9dc21ec00c17fdb0022016f02e6c68865e1eb30076579"} Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.017713 4575 generic.go:334] "Generic (PLEG): container finished" podID="89017745-df9d-4bae-ad06-786a6b653ce8" containerID="c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0" exitCode=0 Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.017955 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.021173 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89017745-df9d-4bae-ad06-786a6b653ce8","Type":"ContainerDied","Data":"c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0"} Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.021227 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89017745-df9d-4bae-ad06-786a6b653ce8","Type":"ContainerDied","Data":"dbb201289bea12734054ef5d370e316daf3fb6c76bda22f46d54827d6792995b"} Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.021251 4575 scope.go:117] "RemoveContainer" containerID="bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.123802 4575 scope.go:117] "RemoveContainer" containerID="c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.182827 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.239170 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.283513 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:06 crc kubenswrapper[4575]: E1004 04:57:06.284008 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" containerName="ceilometer-notification-agent" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.284026 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" containerName="ceilometer-notification-agent" Oct 04 04:57:06 crc kubenswrapper[4575]: E1004 04:57:06.284102 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" containerName="sg-core" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.284114 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" containerName="sg-core" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.284313 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" containerName="ceilometer-notification-agent" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.284344 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" containerName="sg-core" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.286561 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.298929 4575 scope.go:117] "RemoveContainer" containerID="bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.299204 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.299349 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:57:06 crc kubenswrapper[4575]: E1004 04:57:06.300403 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66\": container with ID starting with bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66 not found: ID does not exist" containerID="bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.300514 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66"} err="failed to get container status \"bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66\": rpc error: code = NotFound desc = could not find container \"bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66\": container with ID starting with bc7bbe7826de0df590e5f66262b5d4ab1abbb01f01969da8b0a12ed84e503f66 not found: ID does not exist" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.300624 4575 scope.go:117] "RemoveContainer" containerID="c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.300307 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:06 crc kubenswrapper[4575]: E1004 04:57:06.308691 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0\": container with ID starting with c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0 not found: ID does not exist" containerID="c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.308743 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0"} err="failed to get container status \"c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0\": rpc error: code = NotFound desc = could not find container \"c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0\": container with ID starting with c21360120a346f9b68d63051a87a218ca32a803fd8b0c8c0b24baa16ab0b09f0 not found: ID does not exist" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.315201 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-run-httpd\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.315706 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-log-httpd\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.315927 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-config-data\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.316099 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-scripts\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.316341 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94gk2\" (UniqueName: \"kubernetes.io/projected/f173c031-dfa8-47f2-a520-3d2de14d63b5-kube-api-access-94gk2\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.316722 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.317302 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.422105 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-scripts\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.422178 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94gk2\" (UniqueName: \"kubernetes.io/projected/f173c031-dfa8-47f2-a520-3d2de14d63b5-kube-api-access-94gk2\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.422254 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.422346 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.422366 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-run-httpd\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.422390 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-log-httpd\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.422427 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-config-data\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.424783 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-run-httpd\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.425526 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-log-httpd\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.433870 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-config-data\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.435240 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.441205 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-scripts\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.441600 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.448539 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94gk2\" (UniqueName: \"kubernetes.io/projected/f173c031-dfa8-47f2-a520-3d2de14d63b5-kube-api-access-94gk2\") pod \"ceilometer-0\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.646948 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.714794 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.837293 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-config\") pod \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.837577 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-combined-ca-bundle\") pod \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.837656 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-272pp\" (UniqueName: \"kubernetes.io/projected/d089c375-6e05-4d74-88e1-d00c8cbddb1f-kube-api-access-272pp\") pod \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.843817 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d089c375-6e05-4d74-88e1-d00c8cbddb1f-kube-api-access-272pp" (OuterVolumeSpecName: "kube-api-access-272pp") pod "d089c375-6e05-4d74-88e1-d00c8cbddb1f" (UID: "d089c375-6e05-4d74-88e1-d00c8cbddb1f"). InnerVolumeSpecName "kube-api-access-272pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:06 crc kubenswrapper[4575]: E1004 04:57:06.865963 4575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-config podName:d089c375-6e05-4d74-88e1-d00c8cbddb1f nodeName:}" failed. No retries permitted until 2025-10-04 04:57:07.365939107 +0000 UTC m=+1378.694497921 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-config") pod "d089c375-6e05-4d74-88e1-d00c8cbddb1f" (UID: "d089c375-6e05-4d74-88e1-d00c8cbddb1f") : error deleting /var/lib/kubelet/pods/d089c375-6e05-4d74-88e1-d00c8cbddb1f/volume-subpaths: remove /var/lib/kubelet/pods/d089c375-6e05-4d74-88e1-d00c8cbddb1f/volume-subpaths: no such file or directory Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.869214 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d089c375-6e05-4d74-88e1-d00c8cbddb1f" (UID: "d089c375-6e05-4d74-88e1-d00c8cbddb1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.940868 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:06 crc kubenswrapper[4575]: I1004 04:57:06.941401 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-272pp\" (UniqueName: \"kubernetes.io/projected/d089c375-6e05-4d74-88e1-d00c8cbddb1f-kube-api-access-272pp\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.159990 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" event={"ID":"60168709-79af-4397-b804-fbe06d4bc6e3","Type":"ContainerStarted","Data":"3230dea0554883a20208ba0e92722774fde3f04a122f3182cc3cf4da62c25391"} Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.163664 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.165651 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-7kr5n" event={"ID":"d089c375-6e05-4d74-88e1-d00c8cbddb1f","Type":"ContainerDied","Data":"c06b6d57dbeae7f85c9ccf06f07cc4af6fa0cbb44fce0b2bda78427bd0245a7b"} Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.165695 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c06b6d57dbeae7f85c9ccf06f07cc4af6fa0cbb44fce0b2bda78427bd0245a7b" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.165774 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-7kr5n" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.246435 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.349473 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" podStartSLOduration=3.349446441 podStartE2EDuration="3.349446441s" podCreationTimestamp="2025-10-04 04:57:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:07.229127793 +0000 UTC m=+1378.557686627" watchObservedRunningTime="2025-10-04 04:57:07.349446441 +0000 UTC m=+1378.678005255" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.384992 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89017745-df9d-4bae-ad06-786a6b653ce8" path="/var/lib/kubelet/pods/89017745-df9d-4bae-ad06-786a6b653ce8/volumes" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.464743 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-config\") pod \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\" (UID: \"d089c375-6e05-4d74-88e1-d00c8cbddb1f\") " Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.485394 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-config" (OuterVolumeSpecName: "config") pod "d089c375-6e05-4d74-88e1-d00c8cbddb1f" (UID: "d089c375-6e05-4d74-88e1-d00c8cbddb1f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.551195 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-kcbkv"] Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.567916 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d089c375-6e05-4d74-88e1-d00c8cbddb1f-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.672182 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-fnb6x"] Oct 04 04:57:07 crc kubenswrapper[4575]: E1004 04:57:07.672575 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d089c375-6e05-4d74-88e1-d00c8cbddb1f" containerName="neutron-db-sync" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.672604 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d089c375-6e05-4d74-88e1-d00c8cbddb1f" containerName="neutron-db-sync" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.672796 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="d089c375-6e05-4d74-88e1-d00c8cbddb1f" containerName="neutron-db-sync" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.673707 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.717729 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-fnb6x"] Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.770969 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-config\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.771068 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.771098 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np8nn\" (UniqueName: \"kubernetes.io/projected/b7b66914-2f2d-4068-ae3e-2b742543d07d-kube-api-access-np8nn\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.771120 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.771177 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.771208 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.801656 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6f58cd86dd-q284t"] Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.803310 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.812205 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.812540 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.813375 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.813842 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rxnmx" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.819714 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.841305 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f58cd86dd-q284t"] Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875575 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875659 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np8nn\" (UniqueName: \"kubernetes.io/projected/b7b66914-2f2d-4068-ae3e-2b742543d07d-kube-api-access-np8nn\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875687 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875751 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-combined-ca-bundle\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875776 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875807 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875869 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-httpd-config\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875886 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-ovndb-tls-certs\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875902 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klgj5\" (UniqueName: \"kubernetes.io/projected/86728208-cac4-4663-b3c6-d31786277494-kube-api-access-klgj5\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875927 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-config\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.875947 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-config\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.877030 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-config\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.877540 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.878430 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.891688 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.896337 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.951805 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np8nn\" (UniqueName: \"kubernetes.io/projected/b7b66914-2f2d-4068-ae3e-2b742543d07d-kube-api-access-np8nn\") pod \"dnsmasq-dns-6bb4fc677f-fnb6x\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.988491 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-httpd-config\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.988535 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-ovndb-tls-certs\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.988557 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klgj5\" (UniqueName: \"kubernetes.io/projected/86728208-cac4-4663-b3c6-d31786277494-kube-api-access-klgj5\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.988610 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-config\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:07 crc kubenswrapper[4575]: I1004 04:57:07.988745 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-combined-ca-bundle\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.003179 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-config\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.003698 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.003743 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-ovndb-tls-certs\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.004417 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-combined-ca-bundle\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.023065 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-httpd-config\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.025979 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klgj5\" (UniqueName: \"kubernetes.io/projected/86728208-cac4-4663-b3c6-d31786277494-kube-api-access-klgj5\") pod \"neutron-6f58cd86dd-q284t\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.146321 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.252677 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerStarted","Data":"a3d50e769472dcf76e613e283544583f53a2de958494ac6c2780e382edd7b9ca"} Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.447806 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.447861 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:57:08 crc kubenswrapper[4575]: I1004 04:57:08.727850 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-fnb6x"] Oct 04 04:57:08 crc kubenswrapper[4575]: W1004 04:57:08.849678 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7b66914_2f2d_4068_ae3e_2b742543d07d.slice/crio-21a52c76ebac452ee83b55b1ec49d9bd3148f6e94c41a7777add746520af5368 WatchSource:0}: Error finding container 21a52c76ebac452ee83b55b1ec49d9bd3148f6e94c41a7777add746520af5368: Status 404 returned error can't find the container with id 21a52c76ebac452ee83b55b1ec49d9bd3148f6e94c41a7777add746520af5368 Oct 04 04:57:09 crc kubenswrapper[4575]: I1004 04:57:09.140770 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f58cd86dd-q284t"] Oct 04 04:57:09 crc kubenswrapper[4575]: I1004 04:57:09.369153 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d1075fa9-8a84-493b-85f1-2772ebbc4df2","Type":"ContainerStarted","Data":"5b9edec56b509625b3b15a0f4ab2032bdbd27dfec11f96e1a8bddd129c715524"} Oct 04 04:57:09 crc kubenswrapper[4575]: I1004 04:57:09.369553 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f58cd86dd-q284t" event={"ID":"86728208-cac4-4663-b3c6-d31786277494","Type":"ContainerStarted","Data":"dd11c458d455c7478baf74e0533f5cb79b376e9b5950c86fb5ec9ac763450bed"} Oct 04 04:57:09 crc kubenswrapper[4575]: I1004 04:57:09.369569 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"61d80f56-f92c-4725-8c53-71ad25eee1a5","Type":"ContainerStarted","Data":"d0019de6382b4fe92aef159ea09c53adf952a8beae5d7ace636ed43b665a4551"} Oct 04 04:57:09 crc kubenswrapper[4575]: I1004 04:57:09.369595 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" event={"ID":"b7b66914-2f2d-4068-ae3e-2b742543d07d","Type":"ContainerStarted","Data":"21a52c76ebac452ee83b55b1ec49d9bd3148f6e94c41a7777add746520af5368"} Oct 04 04:57:09 crc kubenswrapper[4575]: I1004 04:57:09.373458 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" podUID="60168709-79af-4397-b804-fbe06d4bc6e3" containerName="dnsmasq-dns" containerID="cri-o://3230dea0554883a20208ba0e92722774fde3f04a122f3182cc3cf4da62c25391" gracePeriod=10 Oct 04 04:57:09 crc kubenswrapper[4575]: I1004 04:57:09.373575 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerStarted","Data":"b8c89b0435addd4bd4f440250f5997ca7d8c0d9c5d4ed6cab82a1a2aec642ff2"} Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.402171 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"61d80f56-f92c-4725-8c53-71ad25eee1a5","Type":"ContainerStarted","Data":"566d86d008adb0fc073faa325ccba0af06fd4cd92ab4d77a99ecab508988d013"} Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.410978 4575 generic.go:334] "Generic (PLEG): container finished" podID="b7b66914-2f2d-4068-ae3e-2b742543d07d" containerID="9be5659b5c55612b221fe9399b00f592fa94fba6eeb0ad41b3f9a5cad387870d" exitCode=0 Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.411169 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" event={"ID":"b7b66914-2f2d-4068-ae3e-2b742543d07d","Type":"ContainerDied","Data":"9be5659b5c55612b221fe9399b00f592fa94fba6eeb0ad41b3f9a5cad387870d"} Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.429979 4575 generic.go:334] "Generic (PLEG): container finished" podID="60168709-79af-4397-b804-fbe06d4bc6e3" containerID="3230dea0554883a20208ba0e92722774fde3f04a122f3182cc3cf4da62c25391" exitCode=0 Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.430038 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" event={"ID":"60168709-79af-4397-b804-fbe06d4bc6e3","Type":"ContainerDied","Data":"3230dea0554883a20208ba0e92722774fde3f04a122f3182cc3cf4da62c25391"} Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.430073 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" event={"ID":"60168709-79af-4397-b804-fbe06d4bc6e3","Type":"ContainerDied","Data":"e0f3c5409003507da7e9dc21ec00c17fdb0022016f02e6c68865e1eb30076579"} Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.430092 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0f3c5409003507da7e9dc21ec00c17fdb0022016f02e6c68865e1eb30076579" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.446167 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.609363 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-sb\") pod \"60168709-79af-4397-b804-fbe06d4bc6e3\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.609855 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-nb\") pod \"60168709-79af-4397-b804-fbe06d4bc6e3\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.609894 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-svc\") pod \"60168709-79af-4397-b804-fbe06d4bc6e3\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.609935 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-swift-storage-0\") pod \"60168709-79af-4397-b804-fbe06d4bc6e3\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.610005 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-config\") pod \"60168709-79af-4397-b804-fbe06d4bc6e3\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.610030 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dbxh\" (UniqueName: \"kubernetes.io/projected/60168709-79af-4397-b804-fbe06d4bc6e3-kube-api-access-5dbxh\") pod \"60168709-79af-4397-b804-fbe06d4bc6e3\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.619823 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60168709-79af-4397-b804-fbe06d4bc6e3-kube-api-access-5dbxh" (OuterVolumeSpecName: "kube-api-access-5dbxh") pod "60168709-79af-4397-b804-fbe06d4bc6e3" (UID: "60168709-79af-4397-b804-fbe06d4bc6e3"). InnerVolumeSpecName "kube-api-access-5dbxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.712608 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60168709-79af-4397-b804-fbe06d4bc6e3" (UID: "60168709-79af-4397-b804-fbe06d4bc6e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.712692 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-svc\") pod \"60168709-79af-4397-b804-fbe06d4bc6e3\" (UID: \"60168709-79af-4397-b804-fbe06d4bc6e3\") " Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.713196 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dbxh\" (UniqueName: \"kubernetes.io/projected/60168709-79af-4397-b804-fbe06d4bc6e3-kube-api-access-5dbxh\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:10 crc kubenswrapper[4575]: W1004 04:57:10.713279 4575 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/60168709-79af-4397-b804-fbe06d4bc6e3/volumes/kubernetes.io~configmap/dns-svc Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.713290 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60168709-79af-4397-b804-fbe06d4bc6e3" (UID: "60168709-79af-4397-b804-fbe06d4bc6e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.805968 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "60168709-79af-4397-b804-fbe06d4bc6e3" (UID: "60168709-79af-4397-b804-fbe06d4bc6e3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.815729 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.815766 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.879417 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "60168709-79af-4397-b804-fbe06d4bc6e3" (UID: "60168709-79af-4397-b804-fbe06d4bc6e3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.883798 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-config" (OuterVolumeSpecName: "config") pod "60168709-79af-4397-b804-fbe06d4bc6e3" (UID: "60168709-79af-4397-b804-fbe06d4bc6e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.889111 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "60168709-79af-4397-b804-fbe06d4bc6e3" (UID: "60168709-79af-4397-b804-fbe06d4bc6e3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.917319 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.917359 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:10 crc kubenswrapper[4575]: I1004 04:57:10.917376 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60168709-79af-4397-b804-fbe06d4bc6e3-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.449143 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d1075fa9-8a84-493b-85f1-2772ebbc4df2","Type":"ContainerStarted","Data":"aa4195f5a9b639cfe100a42a46ca24ff221736e95e1e1e61d4ee632f1e8c05c9"} Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.456833 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f58cd86dd-q284t" event={"ID":"86728208-cac4-4663-b3c6-d31786277494","Type":"ContainerStarted","Data":"faccfd68c3cfda596fe9cd4bdc163906e433497dc979b9cfea9c99b0de34c364"} Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.456883 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f58cd86dd-q284t" event={"ID":"86728208-cac4-4663-b3c6-d31786277494","Type":"ContainerStarted","Data":"dfb7b91a6fce7416fcc271a6caa8a23ec8c948969b91d95494f0fa232fd17a54"} Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.457782 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.480903 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" event={"ID":"b7b66914-2f2d-4068-ae3e-2b742543d07d","Type":"ContainerStarted","Data":"96642340d0958fc56f29cce3f95246afb1a669c35882f75cff8dfe66e7b655f2"} Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.485409 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.6508057 podStartE2EDuration="7.485380302s" podCreationTimestamp="2025-10-04 04:57:04 +0000 UTC" firstStartedPulling="2025-10-04 04:57:05.228679365 +0000 UTC m=+1376.557238179" lastFinishedPulling="2025-10-04 04:57:06.063253967 +0000 UTC m=+1377.391812781" observedRunningTime="2025-10-04 04:57:11.481199821 +0000 UTC m=+1382.809758655" watchObservedRunningTime="2025-10-04 04:57:11.485380302 +0000 UTC m=+1382.813939146" Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.495666 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerName="cinder-api-log" containerID="cri-o://d0019de6382b4fe92aef159ea09c53adf952a8beae5d7ace636ed43b665a4551" gracePeriod=30 Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.495957 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerStarted","Data":"46de58dfd2f1d5d3090aa8db3f9a65578302cc583be0b3b66233aaadd83c4d0b"} Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.496033 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674b76c99f-kcbkv" Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.496711 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.496776 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerName="cinder-api" containerID="cri-o://566d86d008adb0fc073faa325ccba0af06fd4cd92ab4d77a99ecab508988d013" gracePeriod=30 Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.535463 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6f58cd86dd-q284t" podStartSLOduration=4.535443699 podStartE2EDuration="4.535443699s" podCreationTimestamp="2025-10-04 04:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:11.510661823 +0000 UTC m=+1382.839220637" watchObservedRunningTime="2025-10-04 04:57:11.535443699 +0000 UTC m=+1382.864002513" Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.560515 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-kcbkv"] Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.571798 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-kcbkv"] Oct 04 04:57:11 crc kubenswrapper[4575]: I1004 04:57:11.589895 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.5898796619999995 podStartE2EDuration="7.589879662s" podCreationTimestamp="2025-10-04 04:57:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:11.588461631 +0000 UTC m=+1382.917020455" watchObservedRunningTime="2025-10-04 04:57:11.589879662 +0000 UTC m=+1382.918438486" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.249508 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-656b4b6f57-dw59j"] Oct 04 04:57:12 crc kubenswrapper[4575]: E1004 04:57:12.250364 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60168709-79af-4397-b804-fbe06d4bc6e3" containerName="dnsmasq-dns" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.250385 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="60168709-79af-4397-b804-fbe06d4bc6e3" containerName="dnsmasq-dns" Oct 04 04:57:12 crc kubenswrapper[4575]: E1004 04:57:12.250443 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60168709-79af-4397-b804-fbe06d4bc6e3" containerName="init" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.250452 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="60168709-79af-4397-b804-fbe06d4bc6e3" containerName="init" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.250750 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="60168709-79af-4397-b804-fbe06d4bc6e3" containerName="dnsmasq-dns" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.252383 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.259970 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.261124 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.271758 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-75d794bf56-cf4hl" podUID="3262a5ef-127a-4d3f-ba16-1287e6e4bd9f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.271777 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-75d794bf56-cf4hl" podUID="3262a5ef-127a-4d3f-ba16-1287e6e4bd9f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.292439 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-656b4b6f57-dw59j"] Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.354614 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-httpd-config\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.354877 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-internal-tls-certs\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.354949 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-public-tls-certs\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.355000 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh5jj\" (UniqueName: \"kubernetes.io/projected/b22cb08b-c72d-4ae3-b161-3e80aca0a636-kube-api-access-rh5jj\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.355154 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-ovndb-tls-certs\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.355202 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-config\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.369104 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-combined-ca-bundle\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.473632 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-httpd-config\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.473715 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-internal-tls-certs\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.473742 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-public-tls-certs\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.473776 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh5jj\" (UniqueName: \"kubernetes.io/projected/b22cb08b-c72d-4ae3-b161-3e80aca0a636-kube-api-access-rh5jj\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.473862 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-ovndb-tls-certs\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.473891 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-config\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.473921 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-combined-ca-bundle\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.485652 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-httpd-config\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.486789 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-config\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.487782 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-ovndb-tls-certs\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.488326 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-public-tls-certs\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.489857 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-combined-ca-bundle\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.494464 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b22cb08b-c72d-4ae3-b161-3e80aca0a636-internal-tls-certs\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.516989 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh5jj\" (UniqueName: \"kubernetes.io/projected/b22cb08b-c72d-4ae3-b161-3e80aca0a636-kube-api-access-rh5jj\") pod \"neutron-656b4b6f57-dw59j\" (UID: \"b22cb08b-c72d-4ae3-b161-3e80aca0a636\") " pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.535860 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerStarted","Data":"ed4970b93a68166b880f64885aee773c18bd8a0d62ea8f05feaf0753e6cf8f84"} Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.545315 4575 generic.go:334] "Generic (PLEG): container finished" podID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerID="d0019de6382b4fe92aef159ea09c53adf952a8beae5d7ace636ed43b665a4551" exitCode=143 Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.546895 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"61d80f56-f92c-4725-8c53-71ad25eee1a5","Type":"ContainerDied","Data":"d0019de6382b4fe92aef159ea09c53adf952a8beae5d7ace636ed43b665a4551"} Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.547038 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.580055 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:12 crc kubenswrapper[4575]: I1004 04:57:12.637874 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" podStartSLOduration=5.637857942 podStartE2EDuration="5.637857942s" podCreationTimestamp="2025-10-04 04:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:12.633092985 +0000 UTC m=+1383.961651789" watchObservedRunningTime="2025-10-04 04:57:12.637857942 +0000 UTC m=+1383.966416756" Oct 04 04:57:13 crc kubenswrapper[4575]: I1004 04:57:13.196819 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-75d794bf56-cf4hl" podUID="3262a5ef-127a-4d3f-ba16-1287e6e4bd9f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:57:13 crc kubenswrapper[4575]: I1004 04:57:13.196871 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-75d794bf56-cf4hl" podUID="3262a5ef-127a-4d3f-ba16-1287e6e4bd9f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:57:13 crc kubenswrapper[4575]: I1004 04:57:13.328874 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60168709-79af-4397-b804-fbe06d4bc6e3" path="/var/lib/kubelet/pods/60168709-79af-4397-b804-fbe06d4bc6e3/volumes" Oct 04 04:57:13 crc kubenswrapper[4575]: I1004 04:57:13.380707 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-656b4b6f57-dw59j"] Oct 04 04:57:13 crc kubenswrapper[4575]: I1004 04:57:13.599146 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-656b4b6f57-dw59j" event={"ID":"b22cb08b-c72d-4ae3-b161-3e80aca0a636","Type":"ContainerStarted","Data":"f7204f7173338316a9692568670d6b21c10b936c4f4f31d1f17ab4e29efed6ba"} Oct 04 04:57:14 crc kubenswrapper[4575]: I1004 04:57:14.574747 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 04:57:14 crc kubenswrapper[4575]: I1004 04:57:14.576213 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.158:8080/\": dial tcp 10.217.0.158:8080: connect: connection refused" Oct 04 04:57:14 crc kubenswrapper[4575]: I1004 04:57:14.626935 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerStarted","Data":"3ce2089379f8f466a94ec3be275fd074eb7a30cfd509cd79f6b38c897bff1770"} Oct 04 04:57:14 crc kubenswrapper[4575]: I1004 04:57:14.628237 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:57:14 crc kubenswrapper[4575]: I1004 04:57:14.630305 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-656b4b6f57-dw59j" event={"ID":"b22cb08b-c72d-4ae3-b161-3e80aca0a636","Type":"ContainerStarted","Data":"7f523e8cbf76bf348dcc562f09b7989d0532b22ca485f6959cd0726dec1617ec"} Oct 04 04:57:14 crc kubenswrapper[4575]: I1004 04:57:14.630331 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-656b4b6f57-dw59j" event={"ID":"b22cb08b-c72d-4ae3-b161-3e80aca0a636","Type":"ContainerStarted","Data":"8d63c3d8e312cbf61a2ea5739536d823f477b1122a088c27f88c8b6f3ef4a7d4"} Oct 04 04:57:14 crc kubenswrapper[4575]: I1004 04:57:14.630874 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:14 crc kubenswrapper[4575]: I1004 04:57:14.665635 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.770259077 podStartE2EDuration="8.6656155s" podCreationTimestamp="2025-10-04 04:57:06 +0000 UTC" firstStartedPulling="2025-10-04 04:57:07.338702811 +0000 UTC m=+1378.667261615" lastFinishedPulling="2025-10-04 04:57:13.234059224 +0000 UTC m=+1384.562618038" observedRunningTime="2025-10-04 04:57:14.657342951 +0000 UTC m=+1385.985901785" watchObservedRunningTime="2025-10-04 04:57:14.6656155 +0000 UTC m=+1385.994174314" Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.320829 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-75d794bf56-cf4hl" podUID="3262a5ef-127a-4d3f-ba16-1287e6e4bd9f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.320882 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-75d794bf56-cf4hl" podUID="3262a5ef-127a-4d3f-ba16-1287e6e4bd9f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.381879 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.391857 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75d794bf56-cf4hl" Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.414618 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-656b4b6f57-dw59j" podStartSLOduration=5.414601056 podStartE2EDuration="5.414601056s" podCreationTimestamp="2025-10-04 04:57:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:14.705680808 +0000 UTC m=+1386.034239622" watchObservedRunningTime="2025-10-04 04:57:17.414601056 +0000 UTC m=+1388.743159870" Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.489262 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-765dfcdc6b-ctzxp"] Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.489527 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-765dfcdc6b-ctzxp" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api-log" containerID="cri-o://ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269" gracePeriod=30 Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.489702 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-765dfcdc6b-ctzxp" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api" containerID="cri-o://54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274" gracePeriod=30 Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.681310 4575 generic.go:334] "Generic (PLEG): container finished" podID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerID="ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269" exitCode=143 Oct 04 04:57:17 crc kubenswrapper[4575]: I1004 04:57:17.682382 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-765dfcdc6b-ctzxp" event={"ID":"98eae969-87b5-45fe-ac11-eecaacb4a459","Type":"ContainerDied","Data":"ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269"} Oct 04 04:57:18 crc kubenswrapper[4575]: I1004 04:57:18.004773 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:57:18 crc kubenswrapper[4575]: I1004 04:57:18.099768 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-c74ll"] Oct 04 04:57:18 crc kubenswrapper[4575]: I1004 04:57:18.100050 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" podUID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" containerName="dnsmasq-dns" containerID="cri-o://901c7fae38adf9d32365336addacb1a5aa3bba99ee3462617356897f1a80e3d2" gracePeriod=10 Oct 04 04:57:18 crc kubenswrapper[4575]: I1004 04:57:18.119278 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" podUID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: connect: connection refused" Oct 04 04:57:18 crc kubenswrapper[4575]: I1004 04:57:18.206743 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-75d794bf56-cf4hl" podUID="3262a5ef-127a-4d3f-ba16-1287e6e4bd9f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:57:18 crc kubenswrapper[4575]: I1004 04:57:18.207122 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-75d794bf56-cf4hl" podUID="3262a5ef-127a-4d3f-ba16-1287e6e4bd9f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:57:18 crc kubenswrapper[4575]: I1004 04:57:18.702654 4575 generic.go:334] "Generic (PLEG): container finished" podID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" containerID="901c7fae38adf9d32365336addacb1a5aa3bba99ee3462617356897f1a80e3d2" exitCode=0 Oct 04 04:57:18 crc kubenswrapper[4575]: I1004 04:57:18.702716 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" event={"ID":"1978bad0-ef0f-424e-b8e1-964b99e7b4d6","Type":"ContainerDied","Data":"901c7fae38adf9d32365336addacb1a5aa3bba99ee3462617356897f1a80e3d2"} Oct 04 04:57:18 crc kubenswrapper[4575]: I1004 04:57:18.880707 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.020627 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-config\") pod \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.020822 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-nb\") pod \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.020967 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-swift-storage-0\") pod \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.021108 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-sb\") pod \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.021157 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mq25\" (UniqueName: \"kubernetes.io/projected/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-kube-api-access-2mq25\") pod \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.021191 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-svc\") pod \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\" (UID: \"1978bad0-ef0f-424e-b8e1-964b99e7b4d6\") " Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.048894 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-kube-api-access-2mq25" (OuterVolumeSpecName: "kube-api-access-2mq25") pod "1978bad0-ef0f-424e-b8e1-964b99e7b4d6" (UID: "1978bad0-ef0f-424e-b8e1-964b99e7b4d6"). InnerVolumeSpecName "kube-api-access-2mq25". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.123680 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mq25\" (UniqueName: \"kubernetes.io/projected/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-kube-api-access-2mq25\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.131542 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1978bad0-ef0f-424e-b8e1-964b99e7b4d6" (UID: "1978bad0-ef0f-424e-b8e1-964b99e7b4d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.187126 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1978bad0-ef0f-424e-b8e1-964b99e7b4d6" (UID: "1978bad0-ef0f-424e-b8e1-964b99e7b4d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.194943 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1978bad0-ef0f-424e-b8e1-964b99e7b4d6" (UID: "1978bad0-ef0f-424e-b8e1-964b99e7b4d6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.220092 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-config" (OuterVolumeSpecName: "config") pod "1978bad0-ef0f-424e-b8e1-964b99e7b4d6" (UID: "1978bad0-ef0f-424e-b8e1-964b99e7b4d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.228775 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.229494 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.229612 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.229676 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.261922 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1978bad0-ef0f-424e-b8e1-964b99e7b4d6" (UID: "1978bad0-ef0f-424e-b8e1-964b99e7b4d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.332823 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1978bad0-ef0f-424e-b8e1-964b99e7b4d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.712957 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" event={"ID":"1978bad0-ef0f-424e-b8e1-964b99e7b4d6","Type":"ContainerDied","Data":"8078e3ebc2a3d5d66a3aa221e23d505ab66a75457ae74bbb1b6433aad868d785"} Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.713020 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-c74ll" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.715767 4575 scope.go:117] "RemoveContainer" containerID="901c7fae38adf9d32365336addacb1a5aa3bba99ee3462617356897f1a80e3d2" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.755556 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-c74ll"] Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.820939 4575 scope.go:117] "RemoveContainer" containerID="622aea6defd2febb4fadcbd6b319150cbc85f8e5a032c33f4536a84809933586" Oct 04 04:57:19 crc kubenswrapper[4575]: I1004 04:57:19.834286 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-c74ll"] Oct 04 04:57:20 crc kubenswrapper[4575]: I1004 04:57:20.097600 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:57:20 crc kubenswrapper[4575]: I1004 04:57:20.940871 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-765dfcdc6b-ctzxp" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:50624->10.217.0.156:9311: read: connection reset by peer" Oct 04 04:57:20 crc kubenswrapper[4575]: I1004 04:57:20.940919 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-765dfcdc6b-ctzxp" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:50626->10.217.0.156:9311: read: connection reset by peer" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.342875 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" path="/var/lib/kubelet/pods/1978bad0-ef0f-424e-b8e1-964b99e7b4d6/volumes" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.496777 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.575386 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-combined-ca-bundle\") pod \"98eae969-87b5-45fe-ac11-eecaacb4a459\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.575517 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data\") pod \"98eae969-87b5-45fe-ac11-eecaacb4a459\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.575607 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98eae969-87b5-45fe-ac11-eecaacb4a459-logs\") pod \"98eae969-87b5-45fe-ac11-eecaacb4a459\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.575656 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnl6q\" (UniqueName: \"kubernetes.io/projected/98eae969-87b5-45fe-ac11-eecaacb4a459-kube-api-access-rnl6q\") pod \"98eae969-87b5-45fe-ac11-eecaacb4a459\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.575753 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data-custom\") pod \"98eae969-87b5-45fe-ac11-eecaacb4a459\" (UID: \"98eae969-87b5-45fe-ac11-eecaacb4a459\") " Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.578226 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98eae969-87b5-45fe-ac11-eecaacb4a459-logs" (OuterVolumeSpecName: "logs") pod "98eae969-87b5-45fe-ac11-eecaacb4a459" (UID: "98eae969-87b5-45fe-ac11-eecaacb4a459"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.590861 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98eae969-87b5-45fe-ac11-eecaacb4a459-kube-api-access-rnl6q" (OuterVolumeSpecName: "kube-api-access-rnl6q") pod "98eae969-87b5-45fe-ac11-eecaacb4a459" (UID: "98eae969-87b5-45fe-ac11-eecaacb4a459"). InnerVolumeSpecName "kube-api-access-rnl6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.602934 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "98eae969-87b5-45fe-ac11-eecaacb4a459" (UID: "98eae969-87b5-45fe-ac11-eecaacb4a459"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.628702 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98eae969-87b5-45fe-ac11-eecaacb4a459" (UID: "98eae969-87b5-45fe-ac11-eecaacb4a459"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.680607 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnl6q\" (UniqueName: \"kubernetes.io/projected/98eae969-87b5-45fe-ac11-eecaacb4a459-kube-api-access-rnl6q\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.680650 4575 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.680663 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.680675 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98eae969-87b5-45fe-ac11-eecaacb4a459-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.694104 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data" (OuterVolumeSpecName: "config-data") pod "98eae969-87b5-45fe-ac11-eecaacb4a459" (UID: "98eae969-87b5-45fe-ac11-eecaacb4a459"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.709954 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-864464787f-5jtr5" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.738944 4575 generic.go:334] "Generic (PLEG): container finished" podID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerID="54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274" exitCode=0 Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.738994 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-765dfcdc6b-ctzxp" event={"ID":"98eae969-87b5-45fe-ac11-eecaacb4a459","Type":"ContainerDied","Data":"54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274"} Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.739027 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-765dfcdc6b-ctzxp" event={"ID":"98eae969-87b5-45fe-ac11-eecaacb4a459","Type":"ContainerDied","Data":"8b3a70463419547ce8567baed8bb1306a1ff03f0c4a6b98e2e2d3287736f6630"} Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.739048 4575 scope.go:117] "RemoveContainer" containerID="54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.739166 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-765dfcdc6b-ctzxp" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.783136 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98eae969-87b5-45fe-ac11-eecaacb4a459-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.802304 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-765dfcdc6b-ctzxp"] Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.804970 4575 scope.go:117] "RemoveContainer" containerID="ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.831141 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-765dfcdc6b-ctzxp"] Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.862510 4575 scope.go:117] "RemoveContainer" containerID="54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274" Oct 04 04:57:21 crc kubenswrapper[4575]: E1004 04:57:21.864393 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274\": container with ID starting with 54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274 not found: ID does not exist" containerID="54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.864431 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274"} err="failed to get container status \"54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274\": rpc error: code = NotFound desc = could not find container \"54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274\": container with ID starting with 54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274 not found: ID does not exist" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.864457 4575 scope.go:117] "RemoveContainer" containerID="ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269" Oct 04 04:57:21 crc kubenswrapper[4575]: E1004 04:57:21.866562 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269\": container with ID starting with ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269 not found: ID does not exist" containerID="ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269" Oct 04 04:57:21 crc kubenswrapper[4575]: I1004 04:57:21.866616 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269"} err="failed to get container status \"ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269\": rpc error: code = NotFound desc = could not find container \"ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269\": container with ID starting with ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269 not found: ID does not exist" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.322612 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" path="/var/lib/kubelet/pods/98eae969-87b5-45fe-ac11-eecaacb4a459/volumes" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.815130 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 04:57:23 crc kubenswrapper[4575]: E1004 04:57:23.815573 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" containerName="dnsmasq-dns" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.815600 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" containerName="dnsmasq-dns" Oct 04 04:57:23 crc kubenswrapper[4575]: E1004 04:57:23.815611 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" containerName="init" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.815618 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" containerName="init" Oct 04 04:57:23 crc kubenswrapper[4575]: E1004 04:57:23.815642 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.815648 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api" Oct 04 04:57:23 crc kubenswrapper[4575]: E1004 04:57:23.815678 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api-log" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.815686 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api-log" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.815876 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="1978bad0-ef0f-424e-b8e1-964b99e7b4d6" containerName="dnsmasq-dns" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.815897 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.815914 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="98eae969-87b5-45fe-ac11-eecaacb4a459" containerName="barbican-api-log" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.816535 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.822082 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.822414 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-vk4mn" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.845460 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.852107 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.971411 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z99nc\" (UniqueName: \"kubernetes.io/projected/f4b579e2-4a74-4589-b2c1-37437f91b776-kube-api-access-z99nc\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.971849 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b579e2-4a74-4589-b2c1-37437f91b776-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.972022 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f4b579e2-4a74-4589-b2c1-37437f91b776-openstack-config-secret\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:23 crc kubenswrapper[4575]: I1004 04:57:23.972283 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f4b579e2-4a74-4589-b2c1-37437f91b776-openstack-config\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.074407 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f4b579e2-4a74-4589-b2c1-37437f91b776-openstack-config-secret\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.074565 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f4b579e2-4a74-4589-b2c1-37437f91b776-openstack-config\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.074642 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z99nc\" (UniqueName: \"kubernetes.io/projected/f4b579e2-4a74-4589-b2c1-37437f91b776-kube-api-access-z99nc\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.074706 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b579e2-4a74-4589-b2c1-37437f91b776-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.076416 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f4b579e2-4a74-4589-b2c1-37437f91b776-openstack-config\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.087688 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f4b579e2-4a74-4589-b2c1-37437f91b776-openstack-config-secret\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.096306 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b579e2-4a74-4589-b2c1-37437f91b776-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.113832 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z99nc\" (UniqueName: \"kubernetes.io/projected/f4b579e2-4a74-4589-b2c1-37437f91b776-kube-api-access-z99nc\") pod \"openstackclient\" (UID: \"f4b579e2-4a74-4589-b2c1-37437f91b776\") " pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.210572 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.498015 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.583861 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.636572 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.774054 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="cinder-scheduler" containerID="cri-o://5b9edec56b509625b3b15a0f4ab2032bdbd27dfec11f96e1a8bddd129c715524" gracePeriod=30 Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.774694 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="probe" containerID="cri-o://aa4195f5a9b639cfe100a42a46ca24ff221736e95e1e1e61d4ee632f1e8c05c9" gracePeriod=30 Oct 04 04:57:24 crc kubenswrapper[4575]: I1004 04:57:24.841485 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:57:25 crc kubenswrapper[4575]: I1004 04:57:25.783049 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f4b579e2-4a74-4589-b2c1-37437f91b776","Type":"ContainerStarted","Data":"6784807d18ca0b31c22c11fef8b36550a88a810f8554e6934ca4c1ba26a5f33e"} Oct 04 04:57:26 crc kubenswrapper[4575]: I1004 04:57:26.799959 4575 generic.go:334] "Generic (PLEG): container finished" podID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerID="aa4195f5a9b639cfe100a42a46ca24ff221736e95e1e1e61d4ee632f1e8c05c9" exitCode=0 Oct 04 04:57:26 crc kubenswrapper[4575]: I1004 04:57:26.800316 4575 generic.go:334] "Generic (PLEG): container finished" podID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerID="5b9edec56b509625b3b15a0f4ab2032bdbd27dfec11f96e1a8bddd129c715524" exitCode=0 Oct 04 04:57:26 crc kubenswrapper[4575]: I1004 04:57:26.800040 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d1075fa9-8a84-493b-85f1-2772ebbc4df2","Type":"ContainerDied","Data":"aa4195f5a9b639cfe100a42a46ca24ff221736e95e1e1e61d4ee632f1e8c05c9"} Oct 04 04:57:26 crc kubenswrapper[4575]: I1004 04:57:26.800368 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d1075fa9-8a84-493b-85f1-2772ebbc4df2","Type":"ContainerDied","Data":"5b9edec56b509625b3b15a0f4ab2032bdbd27dfec11f96e1a8bddd129c715524"} Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.555479 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.646313 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-scripts\") pod \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.646413 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-combined-ca-bundle\") pod \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.646473 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data\") pod \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.646560 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltwpk\" (UniqueName: \"kubernetes.io/projected/d1075fa9-8a84-493b-85f1-2772ebbc4df2-kube-api-access-ltwpk\") pod \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.646639 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1075fa9-8a84-493b-85f1-2772ebbc4df2-etc-machine-id\") pod \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.646709 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data-custom\") pod \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\" (UID: \"d1075fa9-8a84-493b-85f1-2772ebbc4df2\") " Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.647611 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1075fa9-8a84-493b-85f1-2772ebbc4df2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d1075fa9-8a84-493b-85f1-2772ebbc4df2" (UID: "d1075fa9-8a84-493b-85f1-2772ebbc4df2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.668474 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1075fa9-8a84-493b-85f1-2772ebbc4df2-kube-api-access-ltwpk" (OuterVolumeSpecName: "kube-api-access-ltwpk") pod "d1075fa9-8a84-493b-85f1-2772ebbc4df2" (UID: "d1075fa9-8a84-493b-85f1-2772ebbc4df2"). InnerVolumeSpecName "kube-api-access-ltwpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.680268 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-scripts" (OuterVolumeSpecName: "scripts") pod "d1075fa9-8a84-493b-85f1-2772ebbc4df2" (UID: "d1075fa9-8a84-493b-85f1-2772ebbc4df2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.682757 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d1075fa9-8a84-493b-85f1-2772ebbc4df2" (UID: "d1075fa9-8a84-493b-85f1-2772ebbc4df2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.749039 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.749090 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltwpk\" (UniqueName: \"kubernetes.io/projected/d1075fa9-8a84-493b-85f1-2772ebbc4df2-kube-api-access-ltwpk\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.749107 4575 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1075fa9-8a84-493b-85f1-2772ebbc4df2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.749118 4575 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.786398 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1075fa9-8a84-493b-85f1-2772ebbc4df2" (UID: "d1075fa9-8a84-493b-85f1-2772ebbc4df2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.828971 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d1075fa9-8a84-493b-85f1-2772ebbc4df2","Type":"ContainerDied","Data":"d64d8211704416b22b33137c519222475759a600bdc9f9f44ecf0604bcd173a6"} Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.829035 4575 scope.go:117] "RemoveContainer" containerID="aa4195f5a9b639cfe100a42a46ca24ff221736e95e1e1e61d4ee632f1e8c05c9" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.829213 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.855164 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.869907 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data" (OuterVolumeSpecName: "config-data") pod "d1075fa9-8a84-493b-85f1-2772ebbc4df2" (UID: "d1075fa9-8a84-493b-85f1-2772ebbc4df2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.915096 4575 scope.go:117] "RemoveContainer" containerID="5b9edec56b509625b3b15a0f4ab2032bdbd27dfec11f96e1a8bddd129c715524" Oct 04 04:57:27 crc kubenswrapper[4575]: I1004 04:57:27.957356 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1075fa9-8a84-493b-85f1-2772ebbc4df2-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.174068 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.217082 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.234837 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:57:28 crc kubenswrapper[4575]: E1004 04:57:28.235398 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="cinder-scheduler" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.235423 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="cinder-scheduler" Oct 04 04:57:28 crc kubenswrapper[4575]: E1004 04:57:28.235479 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="probe" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.235488 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="probe" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.235716 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="probe" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.235745 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" containerName="cinder-scheduler" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.250063 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.250174 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.257108 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.369275 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da88db0d-f53e-47e2-bdd7-0cd88ca93762-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.369349 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dz2f\" (UniqueName: \"kubernetes.io/projected/da88db0d-f53e-47e2-bdd7-0cd88ca93762-kube-api-access-7dz2f\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.369377 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-scripts\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.369499 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.369543 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-config-data\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.369696 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.472230 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.472301 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-config-data\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.472338 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.472422 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da88db0d-f53e-47e2-bdd7-0cd88ca93762-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.472446 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dz2f\" (UniqueName: \"kubernetes.io/projected/da88db0d-f53e-47e2-bdd7-0cd88ca93762-kube-api-access-7dz2f\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.472462 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-scripts\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.473115 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da88db0d-f53e-47e2-bdd7-0cd88ca93762-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.481691 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-scripts\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.482436 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-config-data\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.482614 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.483072 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da88db0d-f53e-47e2-bdd7-0cd88ca93762-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.502115 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dz2f\" (UniqueName: \"kubernetes.io/projected/da88db0d-f53e-47e2-bdd7-0cd88ca93762-kube-api-access-7dz2f\") pod \"cinder-scheduler-0\" (UID: \"da88db0d-f53e-47e2-bdd7-0cd88ca93762\") " pod="openstack/cinder-scheduler-0" Oct 04 04:57:28 crc kubenswrapper[4575]: I1004 04:57:28.573030 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:57:29 crc kubenswrapper[4575]: I1004 04:57:29.182723 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:57:29 crc kubenswrapper[4575]: I1004 04:57:29.354271 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1075fa9-8a84-493b-85f1-2772ebbc4df2" path="/var/lib/kubelet/pods/d1075fa9-8a84-493b-85f1-2772ebbc4df2/volumes" Oct 04 04:57:29 crc kubenswrapper[4575]: I1004 04:57:29.874819 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"da88db0d-f53e-47e2-bdd7-0cd88ca93762","Type":"ContainerStarted","Data":"dac4fc9299eae8420481d36649678b9e348eab31fb2cb7eeb0be5c97c22daef0"} Oct 04 04:57:30 crc kubenswrapper[4575]: I1004 04:57:30.886939 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"da88db0d-f53e-47e2-bdd7-0cd88ca93762","Type":"ContainerStarted","Data":"8f69b8c2fa90d6c9544fbce19733d2f74f8afbcdf3f30007e2e1624f66fbd384"} Oct 04 04:57:31 crc kubenswrapper[4575]: I1004 04:57:31.902373 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"da88db0d-f53e-47e2-bdd7-0cd88ca93762","Type":"ContainerStarted","Data":"d7431e8f47f7e996e7ecf2d84ca007b92270308412c9922e9c3769647a259464"} Oct 04 04:57:31 crc kubenswrapper[4575]: I1004 04:57:31.937410 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.937391438 podStartE2EDuration="3.937391438s" podCreationTimestamp="2025-10-04 04:57:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:31.937156421 +0000 UTC m=+1403.265715245" watchObservedRunningTime="2025-10-04 04:57:31.937391438 +0000 UTC m=+1403.265950252" Oct 04 04:57:33 crc kubenswrapper[4575]: I1004 04:57:33.574058 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 04:57:35 crc kubenswrapper[4575]: W1004 04:57:35.794342 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1075fa9_8a84_493b_85f1_2772ebbc4df2.slice/crio-aa4195f5a9b639cfe100a42a46ca24ff221736e95e1e1e61d4ee632f1e8c05c9.scope WatchSource:0}: Error finding container aa4195f5a9b639cfe100a42a46ca24ff221736e95e1e1e61d4ee632f1e8c05c9: Status 404 returned error can't find the container with id aa4195f5a9b639cfe100a42a46ca24ff221736e95e1e1e61d4ee632f1e8c05c9 Oct 04 04:57:35 crc kubenswrapper[4575]: I1004 04:57:35.952387 4575 generic.go:334] "Generic (PLEG): container finished" podID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerID="fa879635a88d7991c3adb1c644cf46f55b5522a646d4725f1a946e36c99c339d" exitCode=137 Oct 04 04:57:35 crc kubenswrapper[4575]: I1004 04:57:35.952479 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerDied","Data":"fa879635a88d7991c3adb1c644cf46f55b5522a646d4725f1a946e36c99c339d"} Oct 04 04:57:35 crc kubenswrapper[4575]: I1004 04:57:35.962464 4575 generic.go:334] "Generic (PLEG): container finished" podID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerID="d57267105118ef90e3d164b8092e906f1721d7d460bbbe2d30820ae6ca7a010d" exitCode=137 Oct 04 04:57:35 crc kubenswrapper[4575]: I1004 04:57:35.962527 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerDied","Data":"d57267105118ef90e3d164b8092e906f1721d7d460bbbe2d30820ae6ca7a010d"} Oct 04 04:57:36 crc kubenswrapper[4575]: E1004 04:57:36.138205 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1978bad0_ef0f_424e_b8e1_964b99e7b4d6.slice/crio-conmon-901c7fae38adf9d32365336addacb1a5aa3bba99ee3462617356897f1a80e3d2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1978bad0_ef0f_424e_b8e1_964b99e7b4d6.slice/crio-8078e3ebc2a3d5d66a3aa221e23d505ab66a75457ae74bbb1b6433aad868d785\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1075fa9_8a84_493b_85f1_2772ebbc4df2.slice/crio-conmon-aa4195f5a9b639cfe100a42a46ca24ff221736e95e1e1e61d4ee632f1e8c05c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod115cafce_0f37_4b82_8b55_c2bbf7297226.slice/crio-d57267105118ef90e3d164b8092e906f1721d7d460bbbe2d30820ae6ca7a010d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f0573f0_1418_43eb_be82_166fb99ed23d.slice/crio-fa879635a88d7991c3adb1c644cf46f55b5522a646d4725f1a946e36c99c339d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1978bad0_ef0f_424e_b8e1_964b99e7b4d6.slice/crio-901c7fae38adf9d32365336addacb1a5aa3bba99ee3462617356897f1a80e3d2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98eae969_87b5_45fe_ac11_eecaacb4a459.slice/crio-conmon-54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod115cafce_0f37_4b82_8b55_c2bbf7297226.slice/crio-conmon-d57267105118ef90e3d164b8092e906f1721d7d460bbbe2d30820ae6ca7a010d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1075fa9_8a84_493b_85f1_2772ebbc4df2.slice/crio-d64d8211704416b22b33137c519222475759a600bdc9f9f44ecf0604bcd173a6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1075fa9_8a84_493b_85f1_2772ebbc4df2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98eae969_87b5_45fe_ac11_eecaacb4a459.slice/crio-conmon-ef00822e0249a8deb469d5449a49badd4af5cfac3276c9c5609a963b71c29269.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98eae969_87b5_45fe_ac11_eecaacb4a459.slice/crio-54d087bbe7acf8705072862a90104e2e5155c3fd01a0892fb0e67ba6ef85e274.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1978bad0_ef0f_424e_b8e1_964b99e7b4d6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1075fa9_8a84_493b_85f1_2772ebbc4df2.slice/crio-conmon-5b9edec56b509625b3b15a0f4ab2032bdbd27dfec11f96e1a8bddd129c715524.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98eae969_87b5_45fe_ac11_eecaacb4a459.slice/crio-8b3a70463419547ce8567baed8bb1306a1ff03f0c4a6b98e2e2d3287736f6630\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1075fa9_8a84_493b_85f1_2772ebbc4df2.slice/crio-5b9edec56b509625b3b15a0f4ab2032bdbd27dfec11f96e1a8bddd129c715524.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98eae969_87b5_45fe_ac11_eecaacb4a459.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:57:36 crc kubenswrapper[4575]: I1004 04:57:36.692630 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.241343 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-654f8bb69f-d4p2v"] Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.244954 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.252785 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.253643 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.253837 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.269079 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-654f8bb69f-d4p2v"] Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.370911 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b760a20b-b48f-411f-bfb8-34ae6fb9c591-log-httpd\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.371009 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gnvz\" (UniqueName: \"kubernetes.io/projected/b760a20b-b48f-411f-bfb8-34ae6fb9c591-kube-api-access-4gnvz\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.371058 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-public-tls-certs\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.371090 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-internal-tls-certs\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.371134 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b760a20b-b48f-411f-bfb8-34ae6fb9c591-run-httpd\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.371186 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b760a20b-b48f-411f-bfb8-34ae6fb9c591-etc-swift\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.371227 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-config-data\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.371330 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-combined-ca-bundle\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.472726 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-combined-ca-bundle\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.473151 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b760a20b-b48f-411f-bfb8-34ae6fb9c591-log-httpd\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.473177 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gnvz\" (UniqueName: \"kubernetes.io/projected/b760a20b-b48f-411f-bfb8-34ae6fb9c591-kube-api-access-4gnvz\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.473239 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-public-tls-certs\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.473284 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-internal-tls-certs\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.473313 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b760a20b-b48f-411f-bfb8-34ae6fb9c591-run-httpd\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.473374 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b760a20b-b48f-411f-bfb8-34ae6fb9c591-etc-swift\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.473450 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-config-data\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.474272 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b760a20b-b48f-411f-bfb8-34ae6fb9c591-log-httpd\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.474390 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b760a20b-b48f-411f-bfb8-34ae6fb9c591-run-httpd\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.484822 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-combined-ca-bundle\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.485844 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-config-data\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.486463 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-internal-tls-certs\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.486683 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b760a20b-b48f-411f-bfb8-34ae6fb9c591-public-tls-certs\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.501283 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gnvz\" (UniqueName: \"kubernetes.io/projected/b760a20b-b48f-411f-bfb8-34ae6fb9c591-kube-api-access-4gnvz\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.503288 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b760a20b-b48f-411f-bfb8-34ae6fb9c591-etc-swift\") pod \"swift-proxy-654f8bb69f-d4p2v\" (UID: \"b760a20b-b48f-411f-bfb8-34ae6fb9c591\") " pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:37 crc kubenswrapper[4575]: I1004 04:57:37.572277 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:38 crc kubenswrapper[4575]: I1004 04:57:38.159292 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:38 crc kubenswrapper[4575]: I1004 04:57:38.446779 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:57:38 crc kubenswrapper[4575]: I1004 04:57:38.446839 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:57:39 crc kubenswrapper[4575]: I1004 04:57:39.209808 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 04:57:40 crc kubenswrapper[4575]: I1004 04:57:40.361982 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:40 crc kubenswrapper[4575]: I1004 04:57:40.362759 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="ceilometer-central-agent" containerID="cri-o://b8c89b0435addd4bd4f440250f5997ca7d8c0d9c5d4ed6cab82a1a2aec642ff2" gracePeriod=30 Oct 04 04:57:40 crc kubenswrapper[4575]: I1004 04:57:40.363098 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="proxy-httpd" containerID="cri-o://3ce2089379f8f466a94ec3be275fd074eb7a30cfd509cd79f6b38c897bff1770" gracePeriod=30 Oct 04 04:57:40 crc kubenswrapper[4575]: I1004 04:57:40.363217 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="ceilometer-notification-agent" containerID="cri-o://46de58dfd2f1d5d3090aa8db3f9a65578302cc583be0b3b66233aaadd83c4d0b" gracePeriod=30 Oct 04 04:57:40 crc kubenswrapper[4575]: I1004 04:57:40.363269 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="sg-core" containerID="cri-o://ed4970b93a68166b880f64885aee773c18bd8a0d62ea8f05feaf0753e6cf8f84" gracePeriod=30 Oct 04 04:57:41 crc kubenswrapper[4575]: I1004 04:57:41.016993 4575 generic.go:334] "Generic (PLEG): container finished" podID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerID="3ce2089379f8f466a94ec3be275fd074eb7a30cfd509cd79f6b38c897bff1770" exitCode=0 Oct 04 04:57:41 crc kubenswrapper[4575]: I1004 04:57:41.017055 4575 generic.go:334] "Generic (PLEG): container finished" podID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerID="ed4970b93a68166b880f64885aee773c18bd8a0d62ea8f05feaf0753e6cf8f84" exitCode=2 Oct 04 04:57:41 crc kubenswrapper[4575]: I1004 04:57:41.017079 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerDied","Data":"3ce2089379f8f466a94ec3be275fd074eb7a30cfd509cd79f6b38c897bff1770"} Oct 04 04:57:41 crc kubenswrapper[4575]: I1004 04:57:41.017126 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerDied","Data":"ed4970b93a68166b880f64885aee773c18bd8a0d62ea8f05feaf0753e6cf8f84"} Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.031157 4575 generic.go:334] "Generic (PLEG): container finished" podID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerID="566d86d008adb0fc073faa325ccba0af06fd4cd92ab4d77a99ecab508988d013" exitCode=137 Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.031224 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"61d80f56-f92c-4725-8c53-71ad25eee1a5","Type":"ContainerDied","Data":"566d86d008adb0fc073faa325ccba0af06fd4cd92ab4d77a99ecab508988d013"} Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.037277 4575 generic.go:334] "Generic (PLEG): container finished" podID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerID="46de58dfd2f1d5d3090aa8db3f9a65578302cc583be0b3b66233aaadd83c4d0b" exitCode=0 Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.037320 4575 generic.go:334] "Generic (PLEG): container finished" podID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerID="b8c89b0435addd4bd4f440250f5997ca7d8c0d9c5d4ed6cab82a1a2aec642ff2" exitCode=0 Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.037341 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerDied","Data":"46de58dfd2f1d5d3090aa8db3f9a65578302cc583be0b3b66233aaadd83c4d0b"} Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.037368 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerDied","Data":"b8c89b0435addd4bd4f440250f5997ca7d8c0d9c5d4ed6cab82a1a2aec642ff2"} Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.611649 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-656b4b6f57-dw59j" Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.745360 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6f58cd86dd-q284t"] Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.745938 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6f58cd86dd-q284t" podUID="86728208-cac4-4663-b3c6-d31786277494" containerName="neutron-api" containerID="cri-o://dfb7b91a6fce7416fcc271a6caa8a23ec8c948969b91d95494f0fa232fd17a54" gracePeriod=30 Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.746710 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6f58cd86dd-q284t" podUID="86728208-cac4-4663-b3c6-d31786277494" containerName="neutron-httpd" containerID="cri-o://faccfd68c3cfda596fe9cd4bdc163906e433497dc979b9cfea9c99b0de34c364" gracePeriod=30 Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.929563 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.937070 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerName="glance-log" containerID="cri-o://ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f" gracePeriod=30 Oct 04 04:57:42 crc kubenswrapper[4575]: I1004 04:57:42.940937 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerName="glance-httpd" containerID="cri-o://72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1" gracePeriod=30 Oct 04 04:57:43 crc kubenswrapper[4575]: I1004 04:57:43.017213 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/glance-default-internal-api-0" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.151:9292/healthcheck\": EOF" Oct 04 04:57:43 crc kubenswrapper[4575]: E1004 04:57:43.763525 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Oct 04 04:57:43 crc kubenswrapper[4575]: E1004 04:57:43.764674 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5fh66fh664h65bhc7h5f7h594hb8h5cdh55ch5c5hbfh65bh55hf9h555h654h58dh566h54fh56fh546h686h64h86h699h668h689h5h5b7h5ddh565q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z99nc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(f4b579e2-4a74-4589-b2c1-37437f91b776): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:57:43 crc kubenswrapper[4575]: E1004 04:57:43.766054 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="f4b579e2-4a74-4589-b2c1-37437f91b776" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.109246 4575 generic.go:334] "Generic (PLEG): container finished" podID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerID="ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f" exitCode=143 Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.109323 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5","Type":"ContainerDied","Data":"ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f"} Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.133804 4575 generic.go:334] "Generic (PLEG): container finished" podID="86728208-cac4-4663-b3c6-d31786277494" containerID="faccfd68c3cfda596fe9cd4bdc163906e433497dc979b9cfea9c99b0de34c364" exitCode=0 Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.134105 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f58cd86dd-q284t" event={"ID":"86728208-cac4-4663-b3c6-d31786277494","Type":"ContainerDied","Data":"faccfd68c3cfda596fe9cd4bdc163906e433497dc979b9cfea9c99b0de34c364"} Oct 04 04:57:44 crc kubenswrapper[4575]: E1004 04:57:44.140449 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="f4b579e2-4a74-4589-b2c1-37437f91b776" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.377550 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.442707 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-scripts\") pod \"61d80f56-f92c-4725-8c53-71ad25eee1a5\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.442765 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61d80f56-f92c-4725-8c53-71ad25eee1a5-etc-machine-id\") pod \"61d80f56-f92c-4725-8c53-71ad25eee1a5\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.442806 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-combined-ca-bundle\") pod \"61d80f56-f92c-4725-8c53-71ad25eee1a5\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.442839 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data\") pod \"61d80f56-f92c-4725-8c53-71ad25eee1a5\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.442931 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data-custom\") pod \"61d80f56-f92c-4725-8c53-71ad25eee1a5\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.442999 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d80f56-f92c-4725-8c53-71ad25eee1a5-logs\") pod \"61d80f56-f92c-4725-8c53-71ad25eee1a5\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.443017 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stlt6\" (UniqueName: \"kubernetes.io/projected/61d80f56-f92c-4725-8c53-71ad25eee1a5-kube-api-access-stlt6\") pod \"61d80f56-f92c-4725-8c53-71ad25eee1a5\" (UID: \"61d80f56-f92c-4725-8c53-71ad25eee1a5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.456627 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61d80f56-f92c-4725-8c53-71ad25eee1a5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "61d80f56-f92c-4725-8c53-71ad25eee1a5" (UID: "61d80f56-f92c-4725-8c53-71ad25eee1a5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.457167 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61d80f56-f92c-4725-8c53-71ad25eee1a5-logs" (OuterVolumeSpecName: "logs") pod "61d80f56-f92c-4725-8c53-71ad25eee1a5" (UID: "61d80f56-f92c-4725-8c53-71ad25eee1a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.457465 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d80f56-f92c-4725-8c53-71ad25eee1a5-kube-api-access-stlt6" (OuterVolumeSpecName: "kube-api-access-stlt6") pod "61d80f56-f92c-4725-8c53-71ad25eee1a5" (UID: "61d80f56-f92c-4725-8c53-71ad25eee1a5"). InnerVolumeSpecName "kube-api-access-stlt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.465920 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "61d80f56-f92c-4725-8c53-71ad25eee1a5" (UID: "61d80f56-f92c-4725-8c53-71ad25eee1a5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.480890 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.484744 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-scripts" (OuterVolumeSpecName: "scripts") pod "61d80f56-f92c-4725-8c53-71ad25eee1a5" (UID: "61d80f56-f92c-4725-8c53-71ad25eee1a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.529703 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61d80f56-f92c-4725-8c53-71ad25eee1a5" (UID: "61d80f56-f92c-4725-8c53-71ad25eee1a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.544793 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-sg-core-conf-yaml\") pod \"f173c031-dfa8-47f2-a520-3d2de14d63b5\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.544859 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-run-httpd\") pod \"f173c031-dfa8-47f2-a520-3d2de14d63b5\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.544977 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-scripts\") pod \"f173c031-dfa8-47f2-a520-3d2de14d63b5\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545003 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-log-httpd\") pod \"f173c031-dfa8-47f2-a520-3d2de14d63b5\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545037 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94gk2\" (UniqueName: \"kubernetes.io/projected/f173c031-dfa8-47f2-a520-3d2de14d63b5-kube-api-access-94gk2\") pod \"f173c031-dfa8-47f2-a520-3d2de14d63b5\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545094 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-combined-ca-bundle\") pod \"f173c031-dfa8-47f2-a520-3d2de14d63b5\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545195 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-config-data\") pod \"f173c031-dfa8-47f2-a520-3d2de14d63b5\" (UID: \"f173c031-dfa8-47f2-a520-3d2de14d63b5\") " Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545664 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d80f56-f92c-4725-8c53-71ad25eee1a5-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545728 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stlt6\" (UniqueName: \"kubernetes.io/projected/61d80f56-f92c-4725-8c53-71ad25eee1a5-kube-api-access-stlt6\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545740 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545748 4575 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/61d80f56-f92c-4725-8c53-71ad25eee1a5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545759 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.545768 4575 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.566233 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f173c031-dfa8-47f2-a520-3d2de14d63b5" (UID: "f173c031-dfa8-47f2-a520-3d2de14d63b5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.567836 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f173c031-dfa8-47f2-a520-3d2de14d63b5" (UID: "f173c031-dfa8-47f2-a520-3d2de14d63b5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.598041 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f173c031-dfa8-47f2-a520-3d2de14d63b5-kube-api-access-94gk2" (OuterVolumeSpecName: "kube-api-access-94gk2") pod "f173c031-dfa8-47f2-a520-3d2de14d63b5" (UID: "f173c031-dfa8-47f2-a520-3d2de14d63b5"). InnerVolumeSpecName "kube-api-access-94gk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.598375 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-scripts" (OuterVolumeSpecName: "scripts") pod "f173c031-dfa8-47f2-a520-3d2de14d63b5" (UID: "f173c031-dfa8-47f2-a520-3d2de14d63b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.640557 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data" (OuterVolumeSpecName: "config-data") pod "61d80f56-f92c-4725-8c53-71ad25eee1a5" (UID: "61d80f56-f92c-4725-8c53-71ad25eee1a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.650126 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.650155 4575 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.650167 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94gk2\" (UniqueName: \"kubernetes.io/projected/f173c031-dfa8-47f2-a520-3d2de14d63b5-kube-api-access-94gk2\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.650179 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d80f56-f92c-4725-8c53-71ad25eee1a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.650188 4575 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f173c031-dfa8-47f2-a520-3d2de14d63b5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.691736 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f173c031-dfa8-47f2-a520-3d2de14d63b5" (UID: "f173c031-dfa8-47f2-a520-3d2de14d63b5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.754304 4575 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.852882 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f173c031-dfa8-47f2-a520-3d2de14d63b5" (UID: "f173c031-dfa8-47f2-a520-3d2de14d63b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.857611 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.909233 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-config-data" (OuterVolumeSpecName: "config-data") pod "f173c031-dfa8-47f2-a520-3d2de14d63b5" (UID: "f173c031-dfa8-47f2-a520-3d2de14d63b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.928577 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-654f8bb69f-d4p2v"] Oct 04 04:57:44 crc kubenswrapper[4575]: I1004 04:57:44.960057 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f173c031-dfa8-47f2-a520-3d2de14d63b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.176154 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f173c031-dfa8-47f2-a520-3d2de14d63b5","Type":"ContainerDied","Data":"a3d50e769472dcf76e613e283544583f53a2de958494ac6c2780e382edd7b9ca"} Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.176219 4575 scope.go:117] "RemoveContainer" containerID="3ce2089379f8f466a94ec3be275fd074eb7a30cfd509cd79f6b38c897bff1770" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.176386 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.186124 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerStarted","Data":"e66aa0b86f8f6573106bccc2c81186d127b7315b211382f7d35959f05ac6d4fb"} Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.192657 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerStarted","Data":"8d0029179c5592b7aefb48edf15d738cddb00579282fc0b3c8d4eb772010beee"} Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.206435 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"61d80f56-f92c-4725-8c53-71ad25eee1a5","Type":"ContainerDied","Data":"c4961240e0d2d3b9473edf1457f6ad4206062be0ebad6d9bdadababdaaf03d18"} Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.206607 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.212109 4575 scope.go:117] "RemoveContainer" containerID="ed4970b93a68166b880f64885aee773c18bd8a0d62ea8f05feaf0753e6cf8f84" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.220403 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-654f8bb69f-d4p2v" event={"ID":"b760a20b-b48f-411f-bfb8-34ae6fb9c591","Type":"ContainerStarted","Data":"b8cc92a41094cf4eadf8458b03909c4169c5ae7c512afcf8aabd4c68dbfbb3ad"} Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.246634 4575 scope.go:117] "RemoveContainer" containerID="46de58dfd2f1d5d3090aa8db3f9a65578302cc583be0b3b66233aaadd83c4d0b" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.264464 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.275047 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.293578 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:45 crc kubenswrapper[4575]: E1004 04:57:45.294253 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerName="cinder-api-log" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294273 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerName="cinder-api-log" Oct 04 04:57:45 crc kubenswrapper[4575]: E1004 04:57:45.294296 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="proxy-httpd" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294306 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="proxy-httpd" Oct 04 04:57:45 crc kubenswrapper[4575]: E1004 04:57:45.294331 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="sg-core" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294339 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="sg-core" Oct 04 04:57:45 crc kubenswrapper[4575]: E1004 04:57:45.294354 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerName="cinder-api" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294360 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerName="cinder-api" Oct 04 04:57:45 crc kubenswrapper[4575]: E1004 04:57:45.294377 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="ceilometer-central-agent" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294385 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="ceilometer-central-agent" Oct 04 04:57:45 crc kubenswrapper[4575]: E1004 04:57:45.294401 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="ceilometer-notification-agent" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294408 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="ceilometer-notification-agent" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294622 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="sg-core" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294643 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="ceilometer-central-agent" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294656 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="ceilometer-notification-agent" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294669 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" containerName="proxy-httpd" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294683 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerName="cinder-api" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.294691 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d80f56-f92c-4725-8c53-71ad25eee1a5" containerName="cinder-api-log" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.297008 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.309076 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.313070 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.313334 4575 scope.go:117] "RemoveContainer" containerID="b8c89b0435addd4bd4f440250f5997ca7d8c0d9c5d4ed6cab82a1a2aec642ff2" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.400523 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f173c031-dfa8-47f2-a520-3d2de14d63b5" path="/var/lib/kubelet/pods/f173c031-dfa8-47f2-a520-3d2de14d63b5/volumes" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.401271 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.401294 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.401306 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.401320 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.403027 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.408016 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.408349 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.412778 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.414553 4575 scope.go:117] "RemoveContainer" containerID="566d86d008adb0fc073faa325ccba0af06fd4cd92ab4d77a99ecab508988d013" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.445670 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.487784 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-log-httpd\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.488166 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-scripts\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.488281 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jlrw\" (UniqueName: \"kubernetes.io/projected/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-kube-api-access-8jlrw\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.488358 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-run-httpd\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.488455 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.488486 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-config-data\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.488538 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.556907 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.556968 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.558991 4575 scope.go:117] "RemoveContainer" containerID="d0019de6382b4fe92aef159ea09c53adf952a8beae5d7ace636ed43b665a4551" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590184 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590264 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jlrw\" (UniqueName: \"kubernetes.io/projected/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-kube-api-access-8jlrw\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590286 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-config-data\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590321 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-config-data-custom\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590340 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-scripts\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590377 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-logs\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590407 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-run-httpd\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590434 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590459 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590476 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590500 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-config-data\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590606 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590663 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svvjq\" (UniqueName: \"kubernetes.io/projected/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-kube-api-access-svvjq\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590695 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590728 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-log-httpd\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.590756 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-scripts\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.591645 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-run-httpd\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.591683 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-log-httpd\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.598128 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.604484 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-config-data\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.604630 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-scripts\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.605027 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.614131 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jlrw\" (UniqueName: \"kubernetes.io/projected/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-kube-api-access-8jlrw\") pod \"ceilometer-0\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.620004 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.695116 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.695212 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-config-data\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.695238 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-config-data-custom\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.695264 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-scripts\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.695321 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-logs\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.695393 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.695433 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.695506 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.695539 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svvjq\" (UniqueName: \"kubernetes.io/projected/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-kube-api-access-svvjq\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.697185 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-logs\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.697602 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.702163 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.707829 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.714704 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.714774 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.719261 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-config-data-custom\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.725499 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.729443 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-config-data\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.730048 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-scripts\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.731091 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svvjq\" (UniqueName: \"kubernetes.io/projected/8f959c7b-2fd1-4ee3-8138-8cac2e6342af-kube-api-access-svvjq\") pod \"cinder-api-0\" (UID: \"8f959c7b-2fd1-4ee3-8138-8cac2e6342af\") " pod="openstack/cinder-api-0" Oct 04 04:57:45 crc kubenswrapper[4575]: I1004 04:57:45.751390 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:57:46 crc kubenswrapper[4575]: I1004 04:57:46.022639 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:46 crc kubenswrapper[4575]: I1004 04:57:46.234797 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-654f8bb69f-d4p2v" event={"ID":"b760a20b-b48f-411f-bfb8-34ae6fb9c591","Type":"ContainerStarted","Data":"2c3bce5f9e4643be07b668e19f44ebecaf2df677a4de7e9af83d49cdd0914c1b"} Oct 04 04:57:46 crc kubenswrapper[4575]: I1004 04:57:46.234856 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-654f8bb69f-d4p2v" event={"ID":"b760a20b-b48f-411f-bfb8-34ae6fb9c591","Type":"ContainerStarted","Data":"3054162963b3a08f819b44ce9da45ccf042ef60f5ae230496d141f92f2159f93"} Oct 04 04:57:46 crc kubenswrapper[4575]: I1004 04:57:46.234917 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:46 crc kubenswrapper[4575]: I1004 04:57:46.261157 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-654f8bb69f-d4p2v" podStartSLOduration=9.261121927 podStartE2EDuration="9.261121927s" podCreationTimestamp="2025-10-04 04:57:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:46.257915944 +0000 UTC m=+1417.586474778" watchObservedRunningTime="2025-10-04 04:57:46.261121927 +0000 UTC m=+1417.589680741" Oct 04 04:57:46 crc kubenswrapper[4575]: I1004 04:57:46.390227 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:57:46 crc kubenswrapper[4575]: I1004 04:57:46.508549 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:57:46 crc kubenswrapper[4575]: W1004 04:57:46.508550 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f959c7b_2fd1_4ee3_8138_8cac2e6342af.slice/crio-3b6c6d7ac543a0412efecc0852827caeaed08411ce26d666d0605078619c33f6 WatchSource:0}: Error finding container 3b6c6d7ac543a0412efecc0852827caeaed08411ce26d666d0605078619c33f6: Status 404 returned error can't find the container with id 3b6c6d7ac543a0412efecc0852827caeaed08411ce26d666d0605078619c33f6 Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.269835 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8f959c7b-2fd1-4ee3-8138-8cac2e6342af","Type":"ContainerStarted","Data":"6b06095b09d6c73cef59801b62dbc1c513b0d83a89ee05241197c3bd0ef9cbfe"} Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.270441 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8f959c7b-2fd1-4ee3-8138-8cac2e6342af","Type":"ContainerStarted","Data":"3b6c6d7ac543a0412efecc0852827caeaed08411ce26d666d0605078619c33f6"} Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.327190 4575 generic.go:334] "Generic (PLEG): container finished" podID="86728208-cac4-4663-b3c6-d31786277494" containerID="dfb7b91a6fce7416fcc271a6caa8a23ec8c948969b91d95494f0fa232fd17a54" exitCode=0 Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.375630 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d80f56-f92c-4725-8c53-71ad25eee1a5" path="/var/lib/kubelet/pods/61d80f56-f92c-4725-8c53-71ad25eee1a5/volumes" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.376570 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.376628 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f58cd86dd-q284t" event={"ID":"86728208-cac4-4663-b3c6-d31786277494","Type":"ContainerDied","Data":"dfb7b91a6fce7416fcc271a6caa8a23ec8c948969b91d95494f0fa232fd17a54"} Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.376659 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerStarted","Data":"e2492bf6860aea369067d4268aef53f47ffb303b7757ee1efff009dbdf544cfb"} Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.648654 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.758542 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-combined-ca-bundle\") pod \"86728208-cac4-4663-b3c6-d31786277494\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.758659 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-ovndb-tls-certs\") pod \"86728208-cac4-4663-b3c6-d31786277494\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.758792 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-httpd-config\") pod \"86728208-cac4-4663-b3c6-d31786277494\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.758855 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-config\") pod \"86728208-cac4-4663-b3c6-d31786277494\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.758922 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klgj5\" (UniqueName: \"kubernetes.io/projected/86728208-cac4-4663-b3c6-d31786277494-kube-api-access-klgj5\") pod \"86728208-cac4-4663-b3c6-d31786277494\" (UID: \"86728208-cac4-4663-b3c6-d31786277494\") " Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.788942 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "86728208-cac4-4663-b3c6-d31786277494" (UID: "86728208-cac4-4663-b3c6-d31786277494"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.789161 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86728208-cac4-4663-b3c6-d31786277494-kube-api-access-klgj5" (OuterVolumeSpecName: "kube-api-access-klgj5") pod "86728208-cac4-4663-b3c6-d31786277494" (UID: "86728208-cac4-4663-b3c6-d31786277494"). InnerVolumeSpecName "kube-api-access-klgj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.876909 4575 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.877121 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klgj5\" (UniqueName: \"kubernetes.io/projected/86728208-cac4-4663-b3c6-d31786277494-kube-api-access-klgj5\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.924992 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-config" (OuterVolumeSpecName: "config") pod "86728208-cac4-4663-b3c6-d31786277494" (UID: "86728208-cac4-4663-b3c6-d31786277494"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.979006 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.985102 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86728208-cac4-4663-b3c6-d31786277494" (UID: "86728208-cac4-4663-b3c6-d31786277494"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:47 crc kubenswrapper[4575]: I1004 04:57:47.989842 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "86728208-cac4-4663-b3c6-d31786277494" (UID: "86728208-cac4-4663-b3c6-d31786277494"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.080636 4575 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.080704 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86728208-cac4-4663-b3c6-d31786277494-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.370570 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f58cd86dd-q284t" event={"ID":"86728208-cac4-4663-b3c6-d31786277494","Type":"ContainerDied","Data":"dd11c458d455c7478baf74e0533f5cb79b376e9b5950c86fb5ec9ac763450bed"} Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.370981 4575 scope.go:117] "RemoveContainer" containerID="faccfd68c3cfda596fe9cd4bdc163906e433497dc979b9cfea9c99b0de34c364" Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.371180 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f58cd86dd-q284t" Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.388197 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerStarted","Data":"b93e21c137ef4219be4ffd1514bc2fdedcdd821d5d810eeb39360a6a6a9ec3a2"} Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.388249 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerStarted","Data":"e3d306d5f49f1eddbd36cd993b5298291d1581f07d79a9a16404caa6331389bb"} Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.422114 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6f58cd86dd-q284t"] Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.429900 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6f58cd86dd-q284t"] Oct 04 04:57:48 crc kubenswrapper[4575]: I1004 04:57:48.430161 4575 scope.go:117] "RemoveContainer" containerID="dfb7b91a6fce7416fcc271a6caa8a23ec8c948969b91d95494f0fa232fd17a54" Oct 04 04:57:49 crc kubenswrapper[4575]: I1004 04:57:49.348542 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86728208-cac4-4663-b3c6-d31786277494" path="/var/lib/kubelet/pods/86728208-cac4-4663-b3c6-d31786277494/volumes" Oct 04 04:57:49 crc kubenswrapper[4575]: I1004 04:57:49.460615 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerStarted","Data":"6876eb1bd6651975a0be801b9fa85251bdef462a494f05b7c4b10ddc466c7bb0"} Oct 04 04:57:49 crc kubenswrapper[4575]: I1004 04:57:49.467937 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8f959c7b-2fd1-4ee3-8138-8cac2e6342af","Type":"ContainerStarted","Data":"1dc9c08181f3140c457f22ec0eb8bae5d88be54ac4582e12533044ca2cd01e75"} Oct 04 04:57:49 crc kubenswrapper[4575]: I1004 04:57:49.468378 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 04:57:49 crc kubenswrapper[4575]: I1004 04:57:49.491712 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.49169463 podStartE2EDuration="4.49169463s" podCreationTimestamp="2025-10-04 04:57:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:49.487928791 +0000 UTC m=+1420.816487605" watchObservedRunningTime="2025-10-04 04:57:49.49169463 +0000 UTC m=+1420.820253444" Oct 04 04:57:50 crc kubenswrapper[4575]: I1004 04:57:50.497635 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="ceilometer-central-agent" containerID="cri-o://e3d306d5f49f1eddbd36cd993b5298291d1581f07d79a9a16404caa6331389bb" gracePeriod=30 Oct 04 04:57:50 crc kubenswrapper[4575]: I1004 04:57:50.498070 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerStarted","Data":"4fd65a3502b55d965bdcc4eea8594ee6ef2890b1c4b6dfc2151b3b319e282704"} Oct 04 04:57:50 crc kubenswrapper[4575]: I1004 04:57:50.498221 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:57:50 crc kubenswrapper[4575]: I1004 04:57:50.498260 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="proxy-httpd" containerID="cri-o://4fd65a3502b55d965bdcc4eea8594ee6ef2890b1c4b6dfc2151b3b319e282704" gracePeriod=30 Oct 04 04:57:50 crc kubenswrapper[4575]: I1004 04:57:50.498326 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="ceilometer-notification-agent" containerID="cri-o://b93e21c137ef4219be4ffd1514bc2fdedcdd821d5d810eeb39360a6a6a9ec3a2" gracePeriod=30 Oct 04 04:57:50 crc kubenswrapper[4575]: I1004 04:57:50.498363 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="sg-core" containerID="cri-o://6876eb1bd6651975a0be801b9fa85251bdef462a494f05b7c4b10ddc466c7bb0" gracePeriod=30 Oct 04 04:57:50 crc kubenswrapper[4575]: I1004 04:57:50.533888 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.944764885 podStartE2EDuration="5.533854231s" podCreationTimestamp="2025-10-04 04:57:45 +0000 UTC" firstStartedPulling="2025-10-04 04:57:46.427324599 +0000 UTC m=+1417.755883413" lastFinishedPulling="2025-10-04 04:57:50.016413945 +0000 UTC m=+1421.344972759" observedRunningTime="2025-10-04 04:57:50.525119689 +0000 UTC m=+1421.853678513" watchObservedRunningTime="2025-10-04 04:57:50.533854231 +0000 UTC m=+1421.862413065" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.468497 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.514798 4575 generic.go:334] "Generic (PLEG): container finished" podID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerID="6876eb1bd6651975a0be801b9fa85251bdef462a494f05b7c4b10ddc466c7bb0" exitCode=2 Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.515145 4575 generic.go:334] "Generic (PLEG): container finished" podID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerID="b93e21c137ef4219be4ffd1514bc2fdedcdd821d5d810eeb39360a6a6a9ec3a2" exitCode=0 Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.515157 4575 generic.go:334] "Generic (PLEG): container finished" podID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerID="e3d306d5f49f1eddbd36cd993b5298291d1581f07d79a9a16404caa6331389bb" exitCode=0 Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.515206 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerDied","Data":"6876eb1bd6651975a0be801b9fa85251bdef462a494f05b7c4b10ddc466c7bb0"} Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.515236 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerDied","Data":"b93e21c137ef4219be4ffd1514bc2fdedcdd821d5d810eeb39360a6a6a9ec3a2"} Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.515249 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerDied","Data":"e3d306d5f49f1eddbd36cd993b5298291d1581f07d79a9a16404caa6331389bb"} Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.517103 4575 generic.go:334] "Generic (PLEG): container finished" podID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerID="72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1" exitCode=0 Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.517126 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5","Type":"ContainerDied","Data":"72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1"} Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.517142 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5","Type":"ContainerDied","Data":"50b85a761779110bffd0df4738de47dc3164a1e72b0f50f113896a5c6bc51f98"} Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.517158 4575 scope.go:117] "RemoveContainer" containerID="72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.517271 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.547296 4575 scope.go:117] "RemoveContainer" containerID="ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.569618 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-internal-tls-certs\") pod \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.569690 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-scripts\") pod \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.569769 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.569831 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-config-data\") pod \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.569877 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chfg6\" (UniqueName: \"kubernetes.io/projected/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-kube-api-access-chfg6\") pod \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.569962 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-httpd-run\") pod \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.569982 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-combined-ca-bundle\") pod \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.570009 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-logs\") pod \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\" (UID: \"1375b5e3-0250-4fe2-8d8d-e07c4d7afae5\") " Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.570895 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-logs" (OuterVolumeSpecName: "logs") pod "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" (UID: "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.573251 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" (UID: "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.577919 4575 scope.go:117] "RemoveContainer" containerID="72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.579503 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-scripts" (OuterVolumeSpecName: "scripts") pod "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" (UID: "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:51 crc kubenswrapper[4575]: E1004 04:57:51.579802 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1\": container with ID starting with 72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1 not found: ID does not exist" containerID="72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.580056 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1"} err="failed to get container status \"72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1\": rpc error: code = NotFound desc = could not find container \"72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1\": container with ID starting with 72135564a174092cb64dd57512d355b9ae01f4dc7b900ef1157e6d302264ebb1 not found: ID does not exist" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.580083 4575 scope.go:117] "RemoveContainer" containerID="ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f" Oct 04 04:57:51 crc kubenswrapper[4575]: E1004 04:57:51.580799 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f\": container with ID starting with ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f not found: ID does not exist" containerID="ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.580825 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f"} err="failed to get container status \"ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f\": rpc error: code = NotFound desc = could not find container \"ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f\": container with ID starting with ae953d311bb3855b93fb468f2a5673860ca0cbca095835b21ebdb8f2eebc3e3f not found: ID does not exist" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.596758 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" (UID: "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.596803 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-kube-api-access-chfg6" (OuterVolumeSpecName: "kube-api-access-chfg6") pod "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" (UID: "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5"). InnerVolumeSpecName "kube-api-access-chfg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.620181 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" (UID: "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.644399 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" (UID: "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.670062 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-config-data" (OuterVolumeSpecName: "config-data") pod "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" (UID: "1375b5e3-0250-4fe2-8d8d-e07c4d7afae5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.673430 4575 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.673476 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.673492 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chfg6\" (UniqueName: \"kubernetes.io/projected/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-kube-api-access-chfg6\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.673508 4575 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.673522 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.673536 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.673546 4575 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.673556 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.699943 4575 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.775407 4575 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.856906 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.867495 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.882980 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:57:51 crc kubenswrapper[4575]: E1004 04:57:51.883434 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86728208-cac4-4663-b3c6-d31786277494" containerName="neutron-httpd" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.883455 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="86728208-cac4-4663-b3c6-d31786277494" containerName="neutron-httpd" Oct 04 04:57:51 crc kubenswrapper[4575]: E1004 04:57:51.883466 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86728208-cac4-4663-b3c6-d31786277494" containerName="neutron-api" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.883474 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="86728208-cac4-4663-b3c6-d31786277494" containerName="neutron-api" Oct 04 04:57:51 crc kubenswrapper[4575]: E1004 04:57:51.883489 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerName="glance-log" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.883497 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerName="glance-log" Oct 04 04:57:51 crc kubenswrapper[4575]: E1004 04:57:51.883515 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerName="glance-httpd" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.883520 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerName="glance-httpd" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.883723 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="86728208-cac4-4663-b3c6-d31786277494" containerName="neutron-httpd" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.883744 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerName="glance-httpd" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.883760 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" containerName="glance-log" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.883776 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="86728208-cac4-4663-b3c6-d31786277494" containerName="neutron-api" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.884879 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.887331 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.887475 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 04:57:51 crc kubenswrapper[4575]: I1004 04:57:51.990961 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.086046 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c270011a-523c-473b-ba19-db3a657abe99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.086119 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.086183 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.086210 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjnxj\" (UniqueName: \"kubernetes.io/projected/c270011a-523c-473b-ba19-db3a657abe99-kube-api-access-rjnxj\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.086301 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.086327 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c270011a-523c-473b-ba19-db3a657abe99-logs\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.086363 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.086409 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.188282 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c270011a-523c-473b-ba19-db3a657abe99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.188678 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.188721 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.188739 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjnxj\" (UniqueName: \"kubernetes.io/projected/c270011a-523c-473b-ba19-db3a657abe99-kube-api-access-rjnxj\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.188787 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.188806 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c270011a-523c-473b-ba19-db3a657abe99-logs\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.188829 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.188857 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.189457 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.189622 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c270011a-523c-473b-ba19-db3a657abe99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.190125 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c270011a-523c-473b-ba19-db3a657abe99-logs\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.195231 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.196130 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.196560 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.203556 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c270011a-523c-473b-ba19-db3a657abe99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.209995 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjnxj\" (UniqueName: \"kubernetes.io/projected/c270011a-523c-473b-ba19-db3a657abe99-kube-api-access-rjnxj\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.237111 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"c270011a-523c-473b-ba19-db3a657abe99\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.506072 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.593048 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:52 crc kubenswrapper[4575]: I1004 04:57:52.633077 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-654f8bb69f-d4p2v" Oct 04 04:57:53 crc kubenswrapper[4575]: I1004 04:57:53.029581 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:57:53 crc kubenswrapper[4575]: I1004 04:57:53.087837 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:57:53 crc kubenswrapper[4575]: I1004 04:57:53.088133 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="248dffc1-b899-42d0-8591-9b232d821972" containerName="glance-log" containerID="cri-o://7881f5e72ccc072b2c027f6c95b041540b6fda771da724b8c77f85b5f07ebcd5" gracePeriod=30 Oct 04 04:57:53 crc kubenswrapper[4575]: I1004 04:57:53.088650 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="248dffc1-b899-42d0-8591-9b232d821972" containerName="glance-httpd" containerID="cri-o://61770b9da65731707195429d59dbf791815ad68edffb95d4889d2dbc19b49a76" gracePeriod=30 Oct 04 04:57:53 crc kubenswrapper[4575]: I1004 04:57:53.329466 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1375b5e3-0250-4fe2-8d8d-e07c4d7afae5" path="/var/lib/kubelet/pods/1375b5e3-0250-4fe2-8d8d-e07c4d7afae5/volumes" Oct 04 04:57:53 crc kubenswrapper[4575]: I1004 04:57:53.551994 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c270011a-523c-473b-ba19-db3a657abe99","Type":"ContainerStarted","Data":"1172c0c67c8f630131580130b07cdba9438fdd64731d7b3d9fadae9848ce246e"} Oct 04 04:57:53 crc kubenswrapper[4575]: I1004 04:57:53.554689 4575 generic.go:334] "Generic (PLEG): container finished" podID="248dffc1-b899-42d0-8591-9b232d821972" containerID="7881f5e72ccc072b2c027f6c95b041540b6fda771da724b8c77f85b5f07ebcd5" exitCode=143 Oct 04 04:57:53 crc kubenswrapper[4575]: I1004 04:57:53.554786 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"248dffc1-b899-42d0-8591-9b232d821972","Type":"ContainerDied","Data":"7881f5e72ccc072b2c027f6c95b041540b6fda771da724b8c77f85b5f07ebcd5"} Oct 04 04:57:54 crc kubenswrapper[4575]: I1004 04:57:54.570372 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c270011a-523c-473b-ba19-db3a657abe99","Type":"ContainerStarted","Data":"36bcc93972531850e1553742f5e2534fc5775dacb58cceb485a2a97d383cf027"} Oct 04 04:57:55 crc kubenswrapper[4575]: I1004 04:57:55.558615 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:57:55 crc kubenswrapper[4575]: I1004 04:57:55.581047 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c270011a-523c-473b-ba19-db3a657abe99","Type":"ContainerStarted","Data":"acd4759c6ac992d77e43fb610b7f5ba07337ca31f9eee1c2ae0895496eb3c5f1"} Oct 04 04:57:55 crc kubenswrapper[4575]: I1004 04:57:55.607175 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.6071540970000004 podStartE2EDuration="4.607154097s" podCreationTimestamp="2025-10-04 04:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:55.600915657 +0000 UTC m=+1426.929474471" watchObservedRunningTime="2025-10-04 04:57:55.607154097 +0000 UTC m=+1426.935712911" Oct 04 04:57:55 crc kubenswrapper[4575]: I1004 04:57:55.714811 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.624833 4575 generic.go:334] "Generic (PLEG): container finished" podID="248dffc1-b899-42d0-8591-9b232d821972" containerID="61770b9da65731707195429d59dbf791815ad68edffb95d4889d2dbc19b49a76" exitCode=0 Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.625208 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"248dffc1-b899-42d0-8591-9b232d821972","Type":"ContainerDied","Data":"61770b9da65731707195429d59dbf791815ad68edffb95d4889d2dbc19b49a76"} Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.657746 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f4b579e2-4a74-4589-b2c1-37437f91b776","Type":"ContainerStarted","Data":"9943629ce55962f6e7642bdb9023b1a5edca1209a7638cf21074bb91a00dfb72"} Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.810595 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.842564 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.645607213 podStartE2EDuration="33.842546373s" podCreationTimestamp="2025-10-04 04:57:23 +0000 UTC" firstStartedPulling="2025-10-04 04:57:24.864076959 +0000 UTC m=+1396.192635773" lastFinishedPulling="2025-10-04 04:57:56.061016119 +0000 UTC m=+1427.389574933" observedRunningTime="2025-10-04 04:57:56.700574768 +0000 UTC m=+1428.029133602" watchObservedRunningTime="2025-10-04 04:57:56.842546373 +0000 UTC m=+1428.171105187" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.907734 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-httpd-run\") pod \"248dffc1-b899-42d0-8591-9b232d821972\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.907807 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-scripts\") pod \"248dffc1-b899-42d0-8591-9b232d821972\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.907835 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"248dffc1-b899-42d0-8591-9b232d821972\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.907858 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-logs\") pod \"248dffc1-b899-42d0-8591-9b232d821972\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.908318 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-logs" (OuterVolumeSpecName: "logs") pod "248dffc1-b899-42d0-8591-9b232d821972" (UID: "248dffc1-b899-42d0-8591-9b232d821972"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.908791 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "248dffc1-b899-42d0-8591-9b232d821972" (UID: "248dffc1-b899-42d0-8591-9b232d821972"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.908806 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlxn9\" (UniqueName: \"kubernetes.io/projected/248dffc1-b899-42d0-8591-9b232d821972-kube-api-access-xlxn9\") pod \"248dffc1-b899-42d0-8591-9b232d821972\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.908887 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-combined-ca-bundle\") pod \"248dffc1-b899-42d0-8591-9b232d821972\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.908973 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-config-data\") pod \"248dffc1-b899-42d0-8591-9b232d821972\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.909025 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-public-tls-certs\") pod \"248dffc1-b899-42d0-8591-9b232d821972\" (UID: \"248dffc1-b899-42d0-8591-9b232d821972\") " Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.909560 4575 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.909598 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248dffc1-b899-42d0-8591-9b232d821972-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.918014 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/248dffc1-b899-42d0-8591-9b232d821972-kube-api-access-xlxn9" (OuterVolumeSpecName: "kube-api-access-xlxn9") pod "248dffc1-b899-42d0-8591-9b232d821972" (UID: "248dffc1-b899-42d0-8591-9b232d821972"). InnerVolumeSpecName "kube-api-access-xlxn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.932771 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "248dffc1-b899-42d0-8591-9b232d821972" (UID: "248dffc1-b899-42d0-8591-9b232d821972"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.932780 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-scripts" (OuterVolumeSpecName: "scripts") pod "248dffc1-b899-42d0-8591-9b232d821972" (UID: "248dffc1-b899-42d0-8591-9b232d821972"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:56 crc kubenswrapper[4575]: I1004 04:57:56.959034 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "248dffc1-b899-42d0-8591-9b232d821972" (UID: "248dffc1-b899-42d0-8591-9b232d821972"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.012603 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.012641 4575 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.012658 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlxn9\" (UniqueName: \"kubernetes.io/projected/248dffc1-b899-42d0-8591-9b232d821972-kube-api-access-xlxn9\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.012668 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.036383 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "248dffc1-b899-42d0-8591-9b232d821972" (UID: "248dffc1-b899-42d0-8591-9b232d821972"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.063306 4575 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.086737 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-config-data" (OuterVolumeSpecName: "config-data") pod "248dffc1-b899-42d0-8591-9b232d821972" (UID: "248dffc1-b899-42d0-8591-9b232d821972"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.114450 4575 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.114703 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.114779 4575 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248dffc1-b899-42d0-8591-9b232d821972-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.667780 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"248dffc1-b899-42d0-8591-9b232d821972","Type":"ContainerDied","Data":"e30e8d800b1169083994d3220d66e42909f76b725a80516435c4ce3c262b8c0c"} Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.667833 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.668144 4575 scope.go:117] "RemoveContainer" containerID="61770b9da65731707195429d59dbf791815ad68edffb95d4889d2dbc19b49a76" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.700167 4575 scope.go:117] "RemoveContainer" containerID="7881f5e72ccc072b2c027f6c95b041540b6fda771da724b8c77f85b5f07ebcd5" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.713737 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.734028 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.745225 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:57:57 crc kubenswrapper[4575]: E1004 04:57:57.748689 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="248dffc1-b899-42d0-8591-9b232d821972" containerName="glance-log" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.748936 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="248dffc1-b899-42d0-8591-9b232d821972" containerName="glance-log" Oct 04 04:57:57 crc kubenswrapper[4575]: E1004 04:57:57.749065 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="248dffc1-b899-42d0-8591-9b232d821972" containerName="glance-httpd" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.749175 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="248dffc1-b899-42d0-8591-9b232d821972" containerName="glance-httpd" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.749537 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="248dffc1-b899-42d0-8591-9b232d821972" containerName="glance-httpd" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.749667 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="248dffc1-b899-42d0-8591-9b232d821972" containerName="glance-log" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.750933 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.754933 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.755152 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.779757 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.930075 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.930131 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-config-data\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.930161 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9351c479-a1a5-400e-8c8b-358cd9587f8c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.930216 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-scripts\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.930282 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.930417 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.930475 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22mws\" (UniqueName: \"kubernetes.io/projected/9351c479-a1a5-400e-8c8b-358cd9587f8c-kube-api-access-22mws\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:57 crc kubenswrapper[4575]: I1004 04:57:57.930601 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9351c479-a1a5-400e-8c8b-358cd9587f8c-logs\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.031882 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9351c479-a1a5-400e-8c8b-358cd9587f8c-logs\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.032168 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.032258 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-config-data\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.032346 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9351c479-a1a5-400e-8c8b-358cd9587f8c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.032470 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-scripts\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.032685 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.032834 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.032934 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.032499 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9351c479-a1a5-400e-8c8b-358cd9587f8c-logs\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.032879 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9351c479-a1a5-400e-8c8b-358cd9587f8c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.033183 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22mws\" (UniqueName: \"kubernetes.io/projected/9351c479-a1a5-400e-8c8b-358cd9587f8c-kube-api-access-22mws\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.039380 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.065494 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-scripts\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.065840 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-config-data\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.074314 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9351c479-a1a5-400e-8c8b-358cd9587f8c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.077375 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22mws\" (UniqueName: \"kubernetes.io/projected/9351c479-a1a5-400e-8c8b-358cd9587f8c-kube-api-access-22mws\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.085413 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9351c479-a1a5-400e-8c8b-358cd9587f8c\") " pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: I1004 04:57:58.376840 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:57:58 crc kubenswrapper[4575]: W1004 04:57:58.979089 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9351c479_a1a5_400e_8c8b_358cd9587f8c.slice/crio-1d8c24d51448965930badec9d8a956ab7d8e3bf1b3ecde6f843880a48ef1c26f WatchSource:0}: Error finding container 1d8c24d51448965930badec9d8a956ab7d8e3bf1b3ecde6f843880a48ef1c26f: Status 404 returned error can't find the container with id 1d8c24d51448965930badec9d8a956ab7d8e3bf1b3ecde6f843880a48ef1c26f Oct 04 04:57:59 crc kubenswrapper[4575]: I1004 04:57:59.001816 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:57:59 crc kubenswrapper[4575]: I1004 04:57:59.263997 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 04:57:59 crc kubenswrapper[4575]: I1004 04:57:59.370250 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="248dffc1-b899-42d0-8591-9b232d821972" path="/var/lib/kubelet/pods/248dffc1-b899-42d0-8591-9b232d821972/volumes" Oct 04 04:57:59 crc kubenswrapper[4575]: I1004 04:57:59.705299 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9351c479-a1a5-400e-8c8b-358cd9587f8c","Type":"ContainerStarted","Data":"1d8c24d51448965930badec9d8a956ab7d8e3bf1b3ecde6f843880a48ef1c26f"} Oct 04 04:58:00 crc kubenswrapper[4575]: I1004 04:58:00.734401 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9351c479-a1a5-400e-8c8b-358cd9587f8c","Type":"ContainerStarted","Data":"b6ee04ec1e7d960a1802d7cc52bcb962612495531f49c38c4ce6fb88b6bcf3df"} Oct 04 04:58:01 crc kubenswrapper[4575]: I1004 04:58:01.749170 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9351c479-a1a5-400e-8c8b-358cd9587f8c","Type":"ContainerStarted","Data":"d7385ec6f44b16361c1607c69d33ddcc08230508efbdc94a6768fc3e1d288ca1"} Oct 04 04:58:01 crc kubenswrapper[4575]: I1004 04:58:01.778782 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.778761744 podStartE2EDuration="4.778761744s" podCreationTimestamp="2025-10-04 04:57:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:58:01.769009683 +0000 UTC m=+1433.097568507" watchObservedRunningTime="2025-10-04 04:58:01.778761744 +0000 UTC m=+1433.107320558" Oct 04 04:58:02 crc kubenswrapper[4575]: I1004 04:58:02.506638 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:58:02 crc kubenswrapper[4575]: I1004 04:58:02.506711 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:58:02 crc kubenswrapper[4575]: I1004 04:58:02.551359 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:58:02 crc kubenswrapper[4575]: I1004 04:58:02.551699 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:58:02 crc kubenswrapper[4575]: I1004 04:58:02.769219 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:58:02 crc kubenswrapper[4575]: I1004 04:58:02.770577 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:58:04 crc kubenswrapper[4575]: I1004 04:58:04.787884 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:58:04 crc kubenswrapper[4575]: I1004 04:58:04.789054 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:58:05 crc kubenswrapper[4575]: I1004 04:58:05.558407 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:58:05 crc kubenswrapper[4575]: I1004 04:58:05.716268 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:58:06 crc kubenswrapper[4575]: I1004 04:58:06.053885 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:58:06 crc kubenswrapper[4575]: I1004 04:58:06.054027 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:58:06 crc kubenswrapper[4575]: I1004 04:58:06.061754 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.377897 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.378274 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.423494 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.439275 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.445956 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.446022 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.446392 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.447330 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fec7eca9bd28e99b532e638e857f42e7eb7fe8c1e89c0dd52e522352fd86d19d"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.447409 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://fec7eca9bd28e99b532e638e857f42e7eb7fe8c1e89c0dd52e522352fd86d19d" gracePeriod=600 Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.831495 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="fec7eca9bd28e99b532e638e857f42e7eb7fe8c1e89c0dd52e522352fd86d19d" exitCode=0 Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.831567 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"fec7eca9bd28e99b532e638e857f42e7eb7fe8c1e89c0dd52e522352fd86d19d"} Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.831899 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22"} Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.831922 4575 scope.go:117] "RemoveContainer" containerID="3ed1625c85b01a1f2dd9735e9997d0b74d3638c1dc9b14bc0e29139e2f4da8d2" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.832636 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:58:08 crc kubenswrapper[4575]: I1004 04:58:08.832680 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.467338 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-fh56z"] Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.468922 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fh56z" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.479632 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-fh56z"] Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.573565 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-bcghq"] Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.575965 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bcghq" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.589814 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bcghq"] Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.598676 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr9zh\" (UniqueName: \"kubernetes.io/projected/bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b-kube-api-access-vr9zh\") pod \"nova-api-db-create-fh56z\" (UID: \"bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b\") " pod="openstack/nova-api-db-create-fh56z" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.701544 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr9zh\" (UniqueName: \"kubernetes.io/projected/bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b-kube-api-access-vr9zh\") pod \"nova-api-db-create-fh56z\" (UID: \"bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b\") " pod="openstack/nova-api-db-create-fh56z" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.701653 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96d54\" (UniqueName: \"kubernetes.io/projected/99b2d0ef-02b1-4a22-b4c0-475d94076370-kube-api-access-96d54\") pod \"nova-cell0-db-create-bcghq\" (UID: \"99b2d0ef-02b1-4a22-b4c0-475d94076370\") " pod="openstack/nova-cell0-db-create-bcghq" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.743483 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr9zh\" (UniqueName: \"kubernetes.io/projected/bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b-kube-api-access-vr9zh\") pod \"nova-api-db-create-fh56z\" (UID: \"bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b\") " pod="openstack/nova-api-db-create-fh56z" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.793607 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fh56z" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.798622 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-j94zx"] Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.800022 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-j94zx" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.802962 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96d54\" (UniqueName: \"kubernetes.io/projected/99b2d0ef-02b1-4a22-b4c0-475d94076370-kube-api-access-96d54\") pod \"nova-cell0-db-create-bcghq\" (UID: \"99b2d0ef-02b1-4a22-b4c0-475d94076370\") " pod="openstack/nova-cell0-db-create-bcghq" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.849030 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-j94zx"] Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.854701 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96d54\" (UniqueName: \"kubernetes.io/projected/99b2d0ef-02b1-4a22-b4c0-475d94076370-kube-api-access-96d54\") pod \"nova-cell0-db-create-bcghq\" (UID: \"99b2d0ef-02b1-4a22-b4c0-475d94076370\") " pod="openstack/nova-cell0-db-create-bcghq" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.907682 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzlgn\" (UniqueName: \"kubernetes.io/projected/6c3e070b-a4ff-4220-a4f9-f9750733357f-kube-api-access-nzlgn\") pod \"nova-cell1-db-create-j94zx\" (UID: \"6c3e070b-a4ff-4220-a4f9-f9750733357f\") " pod="openstack/nova-cell1-db-create-j94zx" Oct 04 04:58:09 crc kubenswrapper[4575]: I1004 04:58:09.911133 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bcghq" Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.029807 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzlgn\" (UniqueName: \"kubernetes.io/projected/6c3e070b-a4ff-4220-a4f9-f9750733357f-kube-api-access-nzlgn\") pod \"nova-cell1-db-create-j94zx\" (UID: \"6c3e070b-a4ff-4220-a4f9-f9750733357f\") " pod="openstack/nova-cell1-db-create-j94zx" Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.067622 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzlgn\" (UniqueName: \"kubernetes.io/projected/6c3e070b-a4ff-4220-a4f9-f9750733357f-kube-api-access-nzlgn\") pod \"nova-cell1-db-create-j94zx\" (UID: \"6c3e070b-a4ff-4220-a4f9-f9750733357f\") " pod="openstack/nova-cell1-db-create-j94zx" Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.151298 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-j94zx" Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.505476 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bcghq"] Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.535537 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-fh56z"] Oct 04 04:58:10 crc kubenswrapper[4575]: W1004 04:58:10.537131 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99b2d0ef_02b1_4a22_b4c0_475d94076370.slice/crio-c068c2371cacf1c250c761a3c946a181db00113fd822adc5bf3e05f2a113821c WatchSource:0}: Error finding container c068c2371cacf1c250c761a3c946a181db00113fd822adc5bf3e05f2a113821c: Status 404 returned error can't find the container with id c068c2371cacf1c250c761a3c946a181db00113fd822adc5bf3e05f2a113821c Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.860397 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-j94zx"] Oct 04 04:58:10 crc kubenswrapper[4575]: W1004 04:58:10.862932 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c3e070b_a4ff_4220_a4f9_f9750733357f.slice/crio-0b4fef0c923ee90b7a3ca06418b66310ab01b60e25f7612a7d020cac34a21d35 WatchSource:0}: Error finding container 0b4fef0c923ee90b7a3ca06418b66310ab01b60e25f7612a7d020cac34a21d35: Status 404 returned error can't find the container with id 0b4fef0c923ee90b7a3ca06418b66310ab01b60e25f7612a7d020cac34a21d35 Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.928092 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bcghq" event={"ID":"99b2d0ef-02b1-4a22-b4c0-475d94076370","Type":"ContainerStarted","Data":"71ab791e8d7de5acca7be2417dd76e530b128e98457332747480d96154161d1d"} Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.928150 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bcghq" event={"ID":"99b2d0ef-02b1-4a22-b4c0-475d94076370","Type":"ContainerStarted","Data":"c068c2371cacf1c250c761a3c946a181db00113fd822adc5bf3e05f2a113821c"} Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.934680 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-j94zx" event={"ID":"6c3e070b-a4ff-4220-a4f9-f9750733357f","Type":"ContainerStarted","Data":"0b4fef0c923ee90b7a3ca06418b66310ab01b60e25f7612a7d020cac34a21d35"} Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.940904 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fh56z" event={"ID":"bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b","Type":"ContainerStarted","Data":"dd9ec1e13da84c8c60dcc935fca3a6e3979137ceee1d35caf019987f29048439"} Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.941133 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fh56z" event={"ID":"bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b","Type":"ContainerStarted","Data":"7714fceb94f302a854101ff9776ed6c144ba58c873ae8f956907a1d6130fd2de"} Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.952200 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-bcghq" podStartSLOduration=1.952182793 podStartE2EDuration="1.952182793s" podCreationTimestamp="2025-10-04 04:58:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:58:10.948056824 +0000 UTC m=+1442.276615638" watchObservedRunningTime="2025-10-04 04:58:10.952182793 +0000 UTC m=+1442.280741607" Oct 04 04:58:10 crc kubenswrapper[4575]: I1004 04:58:10.967323 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-fh56z" podStartSLOduration=1.9673028289999999 podStartE2EDuration="1.967302829s" podCreationTimestamp="2025-10-04 04:58:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:58:10.964395485 +0000 UTC m=+1442.292954299" watchObservedRunningTime="2025-10-04 04:58:10.967302829 +0000 UTC m=+1442.295861643" Oct 04 04:58:11 crc kubenswrapper[4575]: I1004 04:58:11.952179 4575 generic.go:334] "Generic (PLEG): container finished" podID="6c3e070b-a4ff-4220-a4f9-f9750733357f" containerID="9dbb8fed00babf10448b6214c199fbd0bc1d6dfbd9cad3a5741b30d153dddc0b" exitCode=0 Oct 04 04:58:11 crc kubenswrapper[4575]: I1004 04:58:11.952360 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-j94zx" event={"ID":"6c3e070b-a4ff-4220-a4f9-f9750733357f","Type":"ContainerDied","Data":"9dbb8fed00babf10448b6214c199fbd0bc1d6dfbd9cad3a5741b30d153dddc0b"} Oct 04 04:58:11 crc kubenswrapper[4575]: I1004 04:58:11.955625 4575 generic.go:334] "Generic (PLEG): container finished" podID="bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b" containerID="dd9ec1e13da84c8c60dcc935fca3a6e3979137ceee1d35caf019987f29048439" exitCode=0 Oct 04 04:58:11 crc kubenswrapper[4575]: I1004 04:58:11.955668 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fh56z" event={"ID":"bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b","Type":"ContainerDied","Data":"dd9ec1e13da84c8c60dcc935fca3a6e3979137ceee1d35caf019987f29048439"} Oct 04 04:58:11 crc kubenswrapper[4575]: I1004 04:58:11.958156 4575 generic.go:334] "Generic (PLEG): container finished" podID="99b2d0ef-02b1-4a22-b4c0-475d94076370" containerID="71ab791e8d7de5acca7be2417dd76e530b128e98457332747480d96154161d1d" exitCode=0 Oct 04 04:58:11 crc kubenswrapper[4575]: I1004 04:58:11.958229 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bcghq" event={"ID":"99b2d0ef-02b1-4a22-b4c0-475d94076370","Type":"ContainerDied","Data":"71ab791e8d7de5acca7be2417dd76e530b128e98457332747480d96154161d1d"} Oct 04 04:58:11 crc kubenswrapper[4575]: I1004 04:58:11.987322 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:58:11 crc kubenswrapper[4575]: I1004 04:58:11.987426 4575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:58:12 crc kubenswrapper[4575]: I1004 04:58:12.524010 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.604326 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bcghq" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.611656 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fh56z" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.616829 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-j94zx" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.719810 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzlgn\" (UniqueName: \"kubernetes.io/projected/6c3e070b-a4ff-4220-a4f9-f9750733357f-kube-api-access-nzlgn\") pod \"6c3e070b-a4ff-4220-a4f9-f9750733357f\" (UID: \"6c3e070b-a4ff-4220-a4f9-f9750733357f\") " Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.719896 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96d54\" (UniqueName: \"kubernetes.io/projected/99b2d0ef-02b1-4a22-b4c0-475d94076370-kube-api-access-96d54\") pod \"99b2d0ef-02b1-4a22-b4c0-475d94076370\" (UID: \"99b2d0ef-02b1-4a22-b4c0-475d94076370\") " Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.720050 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr9zh\" (UniqueName: \"kubernetes.io/projected/bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b-kube-api-access-vr9zh\") pod \"bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b\" (UID: \"bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b\") " Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.728246 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b-kube-api-access-vr9zh" (OuterVolumeSpecName: "kube-api-access-vr9zh") pod "bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b" (UID: "bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b"). InnerVolumeSpecName "kube-api-access-vr9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.732970 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99b2d0ef-02b1-4a22-b4c0-475d94076370-kube-api-access-96d54" (OuterVolumeSpecName: "kube-api-access-96d54") pod "99b2d0ef-02b1-4a22-b4c0-475d94076370" (UID: "99b2d0ef-02b1-4a22-b4c0-475d94076370"). InnerVolumeSpecName "kube-api-access-96d54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.741694 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c3e070b-a4ff-4220-a4f9-f9750733357f-kube-api-access-nzlgn" (OuterVolumeSpecName: "kube-api-access-nzlgn") pod "6c3e070b-a4ff-4220-a4f9-f9750733357f" (UID: "6c3e070b-a4ff-4220-a4f9-f9750733357f"). InnerVolumeSpecName "kube-api-access-nzlgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.821809 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr9zh\" (UniqueName: \"kubernetes.io/projected/bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b-kube-api-access-vr9zh\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.821845 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzlgn\" (UniqueName: \"kubernetes.io/projected/6c3e070b-a4ff-4220-a4f9-f9750733357f-kube-api-access-nzlgn\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.821857 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96d54\" (UniqueName: \"kubernetes.io/projected/99b2d0ef-02b1-4a22-b4c0-475d94076370-kube-api-access-96d54\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.978546 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-j94zx" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.978553 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-j94zx" event={"ID":"6c3e070b-a4ff-4220-a4f9-f9750733357f","Type":"ContainerDied","Data":"0b4fef0c923ee90b7a3ca06418b66310ab01b60e25f7612a7d020cac34a21d35"} Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.978756 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b4fef0c923ee90b7a3ca06418b66310ab01b60e25f7612a7d020cac34a21d35" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.980877 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fh56z" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.981690 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fh56z" event={"ID":"bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b","Type":"ContainerDied","Data":"7714fceb94f302a854101ff9776ed6c144ba58c873ae8f956907a1d6130fd2de"} Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.981720 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7714fceb94f302a854101ff9776ed6c144ba58c873ae8f956907a1d6130fd2de" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.983319 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bcghq" event={"ID":"99b2d0ef-02b1-4a22-b4c0-475d94076370","Type":"ContainerDied","Data":"c068c2371cacf1c250c761a3c946a181db00113fd822adc5bf3e05f2a113821c"} Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.983340 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c068c2371cacf1c250c761a3c946a181db00113fd822adc5bf3e05f2a113821c" Oct 04 04:58:13 crc kubenswrapper[4575]: I1004 04:58:13.983387 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bcghq" Oct 04 04:58:15 crc kubenswrapper[4575]: I1004 04:58:15.557288 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:58:15 crc kubenswrapper[4575]: I1004 04:58:15.557922 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:58:15 crc kubenswrapper[4575]: I1004 04:58:15.558676 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"8d0029179c5592b7aefb48edf15d738cddb00579282fc0b3c8d4eb772010beee"} pod="openstack/horizon-b86b879b4-tlf42" containerMessage="Container horizon failed startup probe, will be restarted" Oct 04 04:58:15 crc kubenswrapper[4575]: I1004 04:58:15.558709 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" containerID="cri-o://8d0029179c5592b7aefb48edf15d738cddb00579282fc0b3c8d4eb772010beee" gracePeriod=30 Oct 04 04:58:15 crc kubenswrapper[4575]: I1004 04:58:15.653473 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 04 04:58:15 crc kubenswrapper[4575]: I1004 04:58:15.714776 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:58:15 crc kubenswrapper[4575]: I1004 04:58:15.715126 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:58:15 crc kubenswrapper[4575]: I1004 04:58:15.716122 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"e66aa0b86f8f6573106bccc2c81186d127b7315b211382f7d35959f05ac6d4fb"} pod="openstack/horizon-644bf5cdd4-pwdw9" containerMessage="Container horizon failed startup probe, will be restarted" Oct 04 04:58:15 crc kubenswrapper[4575]: I1004 04:58:15.716317 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" containerID="cri-o://e66aa0b86f8f6573106bccc2c81186d127b7315b211382f7d35959f05ac6d4fb" gracePeriod=30 Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.701431 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-aff9-account-create-42nv4"] Oct 04 04:58:19 crc kubenswrapper[4575]: E1004 04:58:19.703234 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c3e070b-a4ff-4220-a4f9-f9750733357f" containerName="mariadb-database-create" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.703309 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c3e070b-a4ff-4220-a4f9-f9750733357f" containerName="mariadb-database-create" Oct 04 04:58:19 crc kubenswrapper[4575]: E1004 04:58:19.703388 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b" containerName="mariadb-database-create" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.703453 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b" containerName="mariadb-database-create" Oct 04 04:58:19 crc kubenswrapper[4575]: E1004 04:58:19.703549 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b2d0ef-02b1-4a22-b4c0-475d94076370" containerName="mariadb-database-create" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.703635 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b2d0ef-02b1-4a22-b4c0-475d94076370" containerName="mariadb-database-create" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.703872 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b" containerName="mariadb-database-create" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.703940 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c3e070b-a4ff-4220-a4f9-f9750733357f" containerName="mariadb-database-create" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.704010 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b2d0ef-02b1-4a22-b4c0-475d94076370" containerName="mariadb-database-create" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.704645 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aff9-account-create-42nv4" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.712086 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.721504 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-aff9-account-create-42nv4"] Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.846600 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxt7p\" (UniqueName: \"kubernetes.io/projected/de414a31-468d-4a87-9aae-26e3b05702da-kube-api-access-mxt7p\") pod \"nova-api-aff9-account-create-42nv4\" (UID: \"de414a31-468d-4a87-9aae-26e3b05702da\") " pod="openstack/nova-api-aff9-account-create-42nv4" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.902842 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f3a6-account-create-7kb28"] Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.903937 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f3a6-account-create-7kb28" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.909614 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.915742 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f3a6-account-create-7kb28"] Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.948462 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxt7p\" (UniqueName: \"kubernetes.io/projected/de414a31-468d-4a87-9aae-26e3b05702da-kube-api-access-mxt7p\") pod \"nova-api-aff9-account-create-42nv4\" (UID: \"de414a31-468d-4a87-9aae-26e3b05702da\") " pod="openstack/nova-api-aff9-account-create-42nv4" Oct 04 04:58:19 crc kubenswrapper[4575]: I1004 04:58:19.980602 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxt7p\" (UniqueName: \"kubernetes.io/projected/de414a31-468d-4a87-9aae-26e3b05702da-kube-api-access-mxt7p\") pod \"nova-api-aff9-account-create-42nv4\" (UID: \"de414a31-468d-4a87-9aae-26e3b05702da\") " pod="openstack/nova-api-aff9-account-create-42nv4" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.030217 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aff9-account-create-42nv4" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.050425 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhs9p\" (UniqueName: \"kubernetes.io/projected/a72b81d1-ed82-47ea-911f-03c05e5f3d32-kube-api-access-rhs9p\") pod \"nova-cell0-f3a6-account-create-7kb28\" (UID: \"a72b81d1-ed82-47ea-911f-03c05e5f3d32\") " pod="openstack/nova-cell0-f3a6-account-create-7kb28" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.124442 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-3e15-account-create-jcnqg"] Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.147199 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3e15-account-create-jcnqg" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.160608 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-3e15-account-create-jcnqg"] Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.162402 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.164237 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhs9p\" (UniqueName: \"kubernetes.io/projected/a72b81d1-ed82-47ea-911f-03c05e5f3d32-kube-api-access-rhs9p\") pod \"nova-cell0-f3a6-account-create-7kb28\" (UID: \"a72b81d1-ed82-47ea-911f-03c05e5f3d32\") " pod="openstack/nova-cell0-f3a6-account-create-7kb28" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.210276 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhs9p\" (UniqueName: \"kubernetes.io/projected/a72b81d1-ed82-47ea-911f-03c05e5f3d32-kube-api-access-rhs9p\") pod \"nova-cell0-f3a6-account-create-7kb28\" (UID: \"a72b81d1-ed82-47ea-911f-03c05e5f3d32\") " pod="openstack/nova-cell0-f3a6-account-create-7kb28" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.221076 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f3a6-account-create-7kb28" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.279096 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb54p\" (UniqueName: \"kubernetes.io/projected/210722f0-a287-4f59-9fb2-8653ec84310e-kube-api-access-pb54p\") pod \"nova-cell1-3e15-account-create-jcnqg\" (UID: \"210722f0-a287-4f59-9fb2-8653ec84310e\") " pod="openstack/nova-cell1-3e15-account-create-jcnqg" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.382942 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb54p\" (UniqueName: \"kubernetes.io/projected/210722f0-a287-4f59-9fb2-8653ec84310e-kube-api-access-pb54p\") pod \"nova-cell1-3e15-account-create-jcnqg\" (UID: \"210722f0-a287-4f59-9fb2-8653ec84310e\") " pod="openstack/nova-cell1-3e15-account-create-jcnqg" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.404605 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb54p\" (UniqueName: \"kubernetes.io/projected/210722f0-a287-4f59-9fb2-8653ec84310e-kube-api-access-pb54p\") pod \"nova-cell1-3e15-account-create-jcnqg\" (UID: \"210722f0-a287-4f59-9fb2-8653ec84310e\") " pod="openstack/nova-cell1-3e15-account-create-jcnqg" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.490174 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3e15-account-create-jcnqg" Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.717562 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-aff9-account-create-42nv4"] Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.784903 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-3e15-account-create-jcnqg"] Oct 04 04:58:20 crc kubenswrapper[4575]: I1004 04:58:20.852529 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f3a6-account-create-7kb28"] Oct 04 04:58:20 crc kubenswrapper[4575]: W1004 04:58:20.853895 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda72b81d1_ed82_47ea_911f_03c05e5f3d32.slice/crio-2a2deae53d519e9d99b5d4461d6b0c29df9758d0ade77dcd91e04cd809014199 WatchSource:0}: Error finding container 2a2deae53d519e9d99b5d4461d6b0c29df9758d0ade77dcd91e04cd809014199: Status 404 returned error can't find the container with id 2a2deae53d519e9d99b5d4461d6b0c29df9758d0ade77dcd91e04cd809014199 Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.049366 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3e15-account-create-jcnqg" event={"ID":"210722f0-a287-4f59-9fb2-8653ec84310e","Type":"ContainerStarted","Data":"1f374dc69a5553d5309029899bf15a1a16162ddb4f17171c2e9346d729cf8ffa"} Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.049752 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3e15-account-create-jcnqg" event={"ID":"210722f0-a287-4f59-9fb2-8653ec84310e","Type":"ContainerStarted","Data":"9e7bbdbdf75e7d59732b70af0f2be3594a97a668422e879e89261be4585ffc2a"} Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.051828 4575 generic.go:334] "Generic (PLEG): container finished" podID="de414a31-468d-4a87-9aae-26e3b05702da" containerID="1e273f1c6f0ed767b53473126090e37faf54ab31ec2dd34c40461d322eb6ff2e" exitCode=0 Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.051898 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aff9-account-create-42nv4" event={"ID":"de414a31-468d-4a87-9aae-26e3b05702da","Type":"ContainerDied","Data":"1e273f1c6f0ed767b53473126090e37faf54ab31ec2dd34c40461d322eb6ff2e"} Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.051915 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aff9-account-create-42nv4" event={"ID":"de414a31-468d-4a87-9aae-26e3b05702da","Type":"ContainerStarted","Data":"c7debf45380a9880d3fd9f951e90628c9bbcc13e4b0b5003bc716f914e7150a8"} Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.056066 4575 generic.go:334] "Generic (PLEG): container finished" podID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerID="4fd65a3502b55d965bdcc4eea8594ee6ef2890b1c4b6dfc2151b3b319e282704" exitCode=137 Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.056107 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerDied","Data":"4fd65a3502b55d965bdcc4eea8594ee6ef2890b1c4b6dfc2151b3b319e282704"} Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.061001 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f3a6-account-create-7kb28" event={"ID":"a72b81d1-ed82-47ea-911f-03c05e5f3d32","Type":"ContainerStarted","Data":"d849fe92b5f9de62575504a96a7948d0429ca8638ee4d1fbc3925e4cc0ea9a6d"} Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.061026 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f3a6-account-create-7kb28" event={"ID":"a72b81d1-ed82-47ea-911f-03c05e5f3d32","Type":"ContainerStarted","Data":"2a2deae53d519e9d99b5d4461d6b0c29df9758d0ade77dcd91e04cd809014199"} Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.074981 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-3e15-account-create-jcnqg" podStartSLOduration=1.074959817 podStartE2EDuration="1.074959817s" podCreationTimestamp="2025-10-04 04:58:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:58:21.071851107 +0000 UTC m=+1452.400409921" watchObservedRunningTime="2025-10-04 04:58:21.074959817 +0000 UTC m=+1452.403518631" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.105136 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-f3a6-account-create-7kb28" podStartSLOduration=2.105114167 podStartE2EDuration="2.105114167s" podCreationTimestamp="2025-10-04 04:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:58:21.099985389 +0000 UTC m=+1452.428544213" watchObservedRunningTime="2025-10-04 04:58:21.105114167 +0000 UTC m=+1452.433673001" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.428184 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.510049 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-combined-ca-bundle\") pod \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.510092 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-log-httpd\") pod \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.510136 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-run-httpd\") pod \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.510169 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-config-data\") pod \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.510253 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-scripts\") pod \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.510297 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-sg-core-conf-yaml\") pod \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.510459 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jlrw\" (UniqueName: \"kubernetes.io/projected/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-kube-api-access-8jlrw\") pod \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\" (UID: \"84d6c6a0-706a-4ad5-aa66-8cbce1e41971\") " Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.511138 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "84d6c6a0-706a-4ad5-aa66-8cbce1e41971" (UID: "84d6c6a0-706a-4ad5-aa66-8cbce1e41971"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.511577 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "84d6c6a0-706a-4ad5-aa66-8cbce1e41971" (UID: "84d6c6a0-706a-4ad5-aa66-8cbce1e41971"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.519538 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-scripts" (OuterVolumeSpecName: "scripts") pod "84d6c6a0-706a-4ad5-aa66-8cbce1e41971" (UID: "84d6c6a0-706a-4ad5-aa66-8cbce1e41971"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.531046 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-kube-api-access-8jlrw" (OuterVolumeSpecName: "kube-api-access-8jlrw") pod "84d6c6a0-706a-4ad5-aa66-8cbce1e41971" (UID: "84d6c6a0-706a-4ad5-aa66-8cbce1e41971"). InnerVolumeSpecName "kube-api-access-8jlrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.565391 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "84d6c6a0-706a-4ad5-aa66-8cbce1e41971" (UID: "84d6c6a0-706a-4ad5-aa66-8cbce1e41971"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.592378 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84d6c6a0-706a-4ad5-aa66-8cbce1e41971" (UID: "84d6c6a0-706a-4ad5-aa66-8cbce1e41971"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.612415 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.612443 4575 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.612452 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jlrw\" (UniqueName: \"kubernetes.io/projected/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-kube-api-access-8jlrw\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.612462 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.612470 4575 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.612478 4575 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.659982 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-config-data" (OuterVolumeSpecName: "config-data") pod "84d6c6a0-706a-4ad5-aa66-8cbce1e41971" (UID: "84d6c6a0-706a-4ad5-aa66-8cbce1e41971"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4575]: I1004 04:58:21.714333 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d6c6a0-706a-4ad5-aa66-8cbce1e41971-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.072235 4575 generic.go:334] "Generic (PLEG): container finished" podID="a72b81d1-ed82-47ea-911f-03c05e5f3d32" containerID="d849fe92b5f9de62575504a96a7948d0429ca8638ee4d1fbc3925e4cc0ea9a6d" exitCode=0 Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.072288 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f3a6-account-create-7kb28" event={"ID":"a72b81d1-ed82-47ea-911f-03c05e5f3d32","Type":"ContainerDied","Data":"d849fe92b5f9de62575504a96a7948d0429ca8638ee4d1fbc3925e4cc0ea9a6d"} Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.074455 4575 generic.go:334] "Generic (PLEG): container finished" podID="210722f0-a287-4f59-9fb2-8653ec84310e" containerID="1f374dc69a5553d5309029899bf15a1a16162ddb4f17171c2e9346d729cf8ffa" exitCode=0 Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.074536 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3e15-account-create-jcnqg" event={"ID":"210722f0-a287-4f59-9fb2-8653ec84310e","Type":"ContainerDied","Data":"1f374dc69a5553d5309029899bf15a1a16162ddb4f17171c2e9346d729cf8ffa"} Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.081503 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.091831 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84d6c6a0-706a-4ad5-aa66-8cbce1e41971","Type":"ContainerDied","Data":"e2492bf6860aea369067d4268aef53f47ffb303b7757ee1efff009dbdf544cfb"} Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.091890 4575 scope.go:117] "RemoveContainer" containerID="4fd65a3502b55d965bdcc4eea8594ee6ef2890b1c4b6dfc2151b3b319e282704" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.157769 4575 scope.go:117] "RemoveContainer" containerID="6876eb1bd6651975a0be801b9fa85251bdef462a494f05b7c4b10ddc466c7bb0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.175727 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.184103 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.206677 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:22 crc kubenswrapper[4575]: E1004 04:58:22.218002 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="ceilometer-central-agent" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.218043 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="ceilometer-central-agent" Oct 04 04:58:22 crc kubenswrapper[4575]: E1004 04:58:22.218060 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="proxy-httpd" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.218068 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="proxy-httpd" Oct 04 04:58:22 crc kubenswrapper[4575]: E1004 04:58:22.218085 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="ceilometer-notification-agent" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.218093 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="ceilometer-notification-agent" Oct 04 04:58:22 crc kubenswrapper[4575]: E1004 04:58:22.218131 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="sg-core" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.218139 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="sg-core" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.218340 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="sg-core" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.218370 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="ceilometer-central-agent" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.218383 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="ceilometer-notification-agent" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.218401 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" containerName="proxy-httpd" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.225115 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.225639 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.227365 4575 scope.go:117] "RemoveContainer" containerID="b93e21c137ef4219be4ffd1514bc2fdedcdd821d5d810eeb39360a6a6a9ec3a2" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.229219 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.238006 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.275893 4575 scope.go:117] "RemoveContainer" containerID="e3d306d5f49f1eddbd36cd993b5298291d1581f07d79a9a16404caa6331389bb" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.330058 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbsx2\" (UniqueName: \"kubernetes.io/projected/c554dcf8-2717-4e95-809f-e61c318f29ca-kube-api-access-pbsx2\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.330129 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-scripts\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.330157 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.330178 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.330233 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-run-httpd\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.330273 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-log-httpd\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.330305 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-config-data\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.436809 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-scripts\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.437525 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.437611 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.437827 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-run-httpd\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.437960 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-log-httpd\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.438063 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-config-data\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.438269 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbsx2\" (UniqueName: \"kubernetes.io/projected/c554dcf8-2717-4e95-809f-e61c318f29ca-kube-api-access-pbsx2\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.439086 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-run-httpd\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.440392 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-log-httpd\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.452516 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.454304 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-scripts\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.455338 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.459732 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-config-data\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.493265 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbsx2\" (UniqueName: \"kubernetes.io/projected/c554dcf8-2717-4e95-809f-e61c318f29ca-kube-api-access-pbsx2\") pod \"ceilometer-0\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.563194 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.709463 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aff9-account-create-42nv4" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.846562 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxt7p\" (UniqueName: \"kubernetes.io/projected/de414a31-468d-4a87-9aae-26e3b05702da-kube-api-access-mxt7p\") pod \"de414a31-468d-4a87-9aae-26e3b05702da\" (UID: \"de414a31-468d-4a87-9aae-26e3b05702da\") " Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.851668 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de414a31-468d-4a87-9aae-26e3b05702da-kube-api-access-mxt7p" (OuterVolumeSpecName: "kube-api-access-mxt7p") pod "de414a31-468d-4a87-9aae-26e3b05702da" (UID: "de414a31-468d-4a87-9aae-26e3b05702da"). InnerVolumeSpecName "kube-api-access-mxt7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:22 crc kubenswrapper[4575]: I1004 04:58:22.949104 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxt7p\" (UniqueName: \"kubernetes.io/projected/de414a31-468d-4a87-9aae-26e3b05702da-kube-api-access-mxt7p\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.097478 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aff9-account-create-42nv4" event={"ID":"de414a31-468d-4a87-9aae-26e3b05702da","Type":"ContainerDied","Data":"c7debf45380a9880d3fd9f951e90628c9bbcc13e4b0b5003bc716f914e7150a8"} Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.097526 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7debf45380a9880d3fd9f951e90628c9bbcc13e4b0b5003bc716f914e7150a8" Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.097614 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aff9-account-create-42nv4" Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.101768 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.112263 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.326648 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d6c6a0-706a-4ad5-aa66-8cbce1e41971" path="/var/lib/kubelet/pods/84d6c6a0-706a-4ad5-aa66-8cbce1e41971/volumes" Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.519407 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f3a6-account-create-7kb28" Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.667133 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhs9p\" (UniqueName: \"kubernetes.io/projected/a72b81d1-ed82-47ea-911f-03c05e5f3d32-kube-api-access-rhs9p\") pod \"a72b81d1-ed82-47ea-911f-03c05e5f3d32\" (UID: \"a72b81d1-ed82-47ea-911f-03c05e5f3d32\") " Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.672747 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a72b81d1-ed82-47ea-911f-03c05e5f3d32-kube-api-access-rhs9p" (OuterVolumeSpecName: "kube-api-access-rhs9p") pod "a72b81d1-ed82-47ea-911f-03c05e5f3d32" (UID: "a72b81d1-ed82-47ea-911f-03c05e5f3d32"). InnerVolumeSpecName "kube-api-access-rhs9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.676116 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3e15-account-create-jcnqg" Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.770906 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb54p\" (UniqueName: \"kubernetes.io/projected/210722f0-a287-4f59-9fb2-8653ec84310e-kube-api-access-pb54p\") pod \"210722f0-a287-4f59-9fb2-8653ec84310e\" (UID: \"210722f0-a287-4f59-9fb2-8653ec84310e\") " Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.771391 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhs9p\" (UniqueName: \"kubernetes.io/projected/a72b81d1-ed82-47ea-911f-03c05e5f3d32-kube-api-access-rhs9p\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.775774 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210722f0-a287-4f59-9fb2-8653ec84310e-kube-api-access-pb54p" (OuterVolumeSpecName: "kube-api-access-pb54p") pod "210722f0-a287-4f59-9fb2-8653ec84310e" (UID: "210722f0-a287-4f59-9fb2-8653ec84310e"). InnerVolumeSpecName "kube-api-access-pb54p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:23 crc kubenswrapper[4575]: I1004 04:58:23.873810 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb54p\" (UniqueName: \"kubernetes.io/projected/210722f0-a287-4f59-9fb2-8653ec84310e-kube-api-access-pb54p\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:24 crc kubenswrapper[4575]: I1004 04:58:24.112858 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerStarted","Data":"7ab01d41cf423474a2609ea04e1a8ba8cac932e91a8ede97711dc866cb3753e9"} Oct 04 04:58:24 crc kubenswrapper[4575]: I1004 04:58:24.112921 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerStarted","Data":"3e1f474c96a167b189c034d7c9747fccd1a564628300c7eb3a1bf33204e00779"} Oct 04 04:58:24 crc kubenswrapper[4575]: I1004 04:58:24.116338 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3e15-account-create-jcnqg" Oct 04 04:58:24 crc kubenswrapper[4575]: I1004 04:58:24.116382 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3e15-account-create-jcnqg" event={"ID":"210722f0-a287-4f59-9fb2-8653ec84310e","Type":"ContainerDied","Data":"9e7bbdbdf75e7d59732b70af0f2be3594a97a668422e879e89261be4585ffc2a"} Oct 04 04:58:24 crc kubenswrapper[4575]: I1004 04:58:24.116421 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e7bbdbdf75e7d59732b70af0f2be3594a97a668422e879e89261be4585ffc2a" Oct 04 04:58:24 crc kubenswrapper[4575]: I1004 04:58:24.118398 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f3a6-account-create-7kb28" event={"ID":"a72b81d1-ed82-47ea-911f-03c05e5f3d32","Type":"ContainerDied","Data":"2a2deae53d519e9d99b5d4461d6b0c29df9758d0ade77dcd91e04cd809014199"} Oct 04 04:58:24 crc kubenswrapper[4575]: I1004 04:58:24.118418 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a2deae53d519e9d99b5d4461d6b0c29df9758d0ade77dcd91e04cd809014199" Oct 04 04:58:24 crc kubenswrapper[4575]: I1004 04:58:24.118475 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f3a6-account-create-7kb28" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.128672 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerStarted","Data":"ba8f91ab188b7ef89bc5187a283e3172ed18fa55ed1e8f21e7bcf280a10203b7"} Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.170726 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2s2b6"] Oct 04 04:58:25 crc kubenswrapper[4575]: E1004 04:58:25.171322 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="210722f0-a287-4f59-9fb2-8653ec84310e" containerName="mariadb-account-create" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.171400 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="210722f0-a287-4f59-9fb2-8653ec84310e" containerName="mariadb-account-create" Oct 04 04:58:25 crc kubenswrapper[4575]: E1004 04:58:25.171463 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de414a31-468d-4a87-9aae-26e3b05702da" containerName="mariadb-account-create" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.171514 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="de414a31-468d-4a87-9aae-26e3b05702da" containerName="mariadb-account-create" Oct 04 04:58:25 crc kubenswrapper[4575]: E1004 04:58:25.171632 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72b81d1-ed82-47ea-911f-03c05e5f3d32" containerName="mariadb-account-create" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.171694 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72b81d1-ed82-47ea-911f-03c05e5f3d32" containerName="mariadb-account-create" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.172009 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="a72b81d1-ed82-47ea-911f-03c05e5f3d32" containerName="mariadb-account-create" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.172120 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="de414a31-468d-4a87-9aae-26e3b05702da" containerName="mariadb-account-create" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.172190 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="210722f0-a287-4f59-9fb2-8653ec84310e" containerName="mariadb-account-create" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.172810 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.177059 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-mz7g2" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.178191 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.178381 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.208060 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2s2b6"] Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.305792 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-scripts\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.305934 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.305974 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp8bm\" (UniqueName: \"kubernetes.io/projected/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-kube-api-access-rp8bm\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.305991 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-config-data\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.407545 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.407926 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp8bm\" (UniqueName: \"kubernetes.io/projected/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-kube-api-access-rp8bm\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.408036 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-config-data\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.408293 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-scripts\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.412800 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-scripts\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.416667 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.436105 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp8bm\" (UniqueName: \"kubernetes.io/projected/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-kube-api-access-rp8bm\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.438057 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-config-data\") pod \"nova-cell0-conductor-db-sync-2s2b6\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:25 crc kubenswrapper[4575]: I1004 04:58:25.494427 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:26 crc kubenswrapper[4575]: I1004 04:58:26.047571 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2s2b6"] Oct 04 04:58:26 crc kubenswrapper[4575]: I1004 04:58:26.160835 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerStarted","Data":"f1d05f0a6d4c4ef80c6ba9cd1817181af964eab80fd5884781091596e8e65b1e"} Oct 04 04:58:26 crc kubenswrapper[4575]: I1004 04:58:26.167139 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2s2b6" event={"ID":"f8d694b1-ab26-49ff-97c6-4a26ef3a0142","Type":"ContainerStarted","Data":"6b5b7ba26cdac66f56744c60d1a9912b2ff09cc0ea4a7cf135700149bd496c2d"} Oct 04 04:58:26 crc kubenswrapper[4575]: I1004 04:58:26.206877 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:27 crc kubenswrapper[4575]: I1004 04:58:27.183342 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerStarted","Data":"79df2b6454e6cb434758d00e83473834c649a5e7d058cd8e05d1c37befe32508"} Oct 04 04:58:27 crc kubenswrapper[4575]: I1004 04:58:27.183736 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="ceilometer-central-agent" containerID="cri-o://7ab01d41cf423474a2609ea04e1a8ba8cac932e91a8ede97711dc866cb3753e9" gracePeriod=30 Oct 04 04:58:27 crc kubenswrapper[4575]: I1004 04:58:27.183769 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="proxy-httpd" containerID="cri-o://79df2b6454e6cb434758d00e83473834c649a5e7d058cd8e05d1c37befe32508" gracePeriod=30 Oct 04 04:58:27 crc kubenswrapper[4575]: I1004 04:58:27.183785 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="sg-core" containerID="cri-o://f1d05f0a6d4c4ef80c6ba9cd1817181af964eab80fd5884781091596e8e65b1e" gracePeriod=30 Oct 04 04:58:27 crc kubenswrapper[4575]: I1004 04:58:27.183820 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:58:27 crc kubenswrapper[4575]: I1004 04:58:27.183804 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="ceilometer-notification-agent" containerID="cri-o://ba8f91ab188b7ef89bc5187a283e3172ed18fa55ed1e8f21e7bcf280a10203b7" gracePeriod=30 Oct 04 04:58:27 crc kubenswrapper[4575]: I1004 04:58:27.213180 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.058361587 podStartE2EDuration="5.213160801s" podCreationTimestamp="2025-10-04 04:58:22 +0000 UTC" firstStartedPulling="2025-10-04 04:58:23.111967697 +0000 UTC m=+1454.440526511" lastFinishedPulling="2025-10-04 04:58:26.266766911 +0000 UTC m=+1457.595325725" observedRunningTime="2025-10-04 04:58:27.212062929 +0000 UTC m=+1458.540621743" watchObservedRunningTime="2025-10-04 04:58:27.213160801 +0000 UTC m=+1458.541719615" Oct 04 04:58:28 crc kubenswrapper[4575]: I1004 04:58:28.207700 4575 generic.go:334] "Generic (PLEG): container finished" podID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerID="79df2b6454e6cb434758d00e83473834c649a5e7d058cd8e05d1c37befe32508" exitCode=0 Oct 04 04:58:28 crc kubenswrapper[4575]: I1004 04:58:28.208268 4575 generic.go:334] "Generic (PLEG): container finished" podID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerID="f1d05f0a6d4c4ef80c6ba9cd1817181af964eab80fd5884781091596e8e65b1e" exitCode=2 Oct 04 04:58:28 crc kubenswrapper[4575]: I1004 04:58:28.207794 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerDied","Data":"79df2b6454e6cb434758d00e83473834c649a5e7d058cd8e05d1c37befe32508"} Oct 04 04:58:28 crc kubenswrapper[4575]: I1004 04:58:28.209295 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerDied","Data":"f1d05f0a6d4c4ef80c6ba9cd1817181af964eab80fd5884781091596e8e65b1e"} Oct 04 04:58:35 crc kubenswrapper[4575]: I1004 04:58:35.293351 4575 generic.go:334] "Generic (PLEG): container finished" podID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerID="7ab01d41cf423474a2609ea04e1a8ba8cac932e91a8ede97711dc866cb3753e9" exitCode=0 Oct 04 04:58:35 crc kubenswrapper[4575]: I1004 04:58:35.295334 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerDied","Data":"7ab01d41cf423474a2609ea04e1a8ba8cac932e91a8ede97711dc866cb3753e9"} Oct 04 04:58:36 crc kubenswrapper[4575]: I1004 04:58:36.304123 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2s2b6" event={"ID":"f8d694b1-ab26-49ff-97c6-4a26ef3a0142","Type":"ContainerStarted","Data":"92ff953ac58e7b34212b6fb3f6650f4156d0273132432a20d45b37fd75280541"} Oct 04 04:58:36 crc kubenswrapper[4575]: I1004 04:58:36.322391 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-2s2b6" podStartSLOduration=2.058454048 podStartE2EDuration="11.322372788s" podCreationTimestamp="2025-10-04 04:58:25 +0000 UTC" firstStartedPulling="2025-10-04 04:58:26.055144196 +0000 UTC m=+1457.383703020" lastFinishedPulling="2025-10-04 04:58:35.319062946 +0000 UTC m=+1466.647621760" observedRunningTime="2025-10-04 04:58:36.32037116 +0000 UTC m=+1467.648929984" watchObservedRunningTime="2025-10-04 04:58:36.322372788 +0000 UTC m=+1467.650931602" Oct 04 04:58:46 crc kubenswrapper[4575]: I1004 04:58:46.393795 4575 generic.go:334] "Generic (PLEG): container finished" podID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerID="e66aa0b86f8f6573106bccc2c81186d127b7315b211382f7d35959f05ac6d4fb" exitCode=137 Oct 04 04:58:46 crc kubenswrapper[4575]: I1004 04:58:46.393855 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerDied","Data":"e66aa0b86f8f6573106bccc2c81186d127b7315b211382f7d35959f05ac6d4fb"} Oct 04 04:58:46 crc kubenswrapper[4575]: I1004 04:58:46.395621 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerStarted","Data":"ec938e12e0dcd7b18a29f344980b6af8ff56070e067b69f96ecbd149ba683cb0"} Oct 04 04:58:46 crc kubenswrapper[4575]: I1004 04:58:46.395702 4575 scope.go:117] "RemoveContainer" containerID="fa879635a88d7991c3adb1c644cf46f55b5522a646d4725f1a946e36c99c339d" Oct 04 04:58:46 crc kubenswrapper[4575]: I1004 04:58:46.398720 4575 generic.go:334] "Generic (PLEG): container finished" podID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerID="8d0029179c5592b7aefb48edf15d738cddb00579282fc0b3c8d4eb772010beee" exitCode=137 Oct 04 04:58:46 crc kubenswrapper[4575]: I1004 04:58:46.398793 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerDied","Data":"8d0029179c5592b7aefb48edf15d738cddb00579282fc0b3c8d4eb772010beee"} Oct 04 04:58:46 crc kubenswrapper[4575]: I1004 04:58:46.398837 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerStarted","Data":"376a8e8470b9959c733967cacb7a6d6b2b157401d6330173cf7a8725128fbf8f"} Oct 04 04:58:46 crc kubenswrapper[4575]: I1004 04:58:46.595882 4575 scope.go:117] "RemoveContainer" containerID="d57267105118ef90e3d164b8092e906f1721d7d460bbbe2d30820ae6ca7a010d" Oct 04 04:58:48 crc kubenswrapper[4575]: E1004 04:58:48.155692 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8d694b1_ab26_49ff_97c6_4a26ef3a0142.slice/crio-conmon-92ff953ac58e7b34212b6fb3f6650f4156d0273132432a20d45b37fd75280541.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:58:48 crc kubenswrapper[4575]: I1004 04:58:48.422160 4575 generic.go:334] "Generic (PLEG): container finished" podID="f8d694b1-ab26-49ff-97c6-4a26ef3a0142" containerID="92ff953ac58e7b34212b6fb3f6650f4156d0273132432a20d45b37fd75280541" exitCode=0 Oct 04 04:58:48 crc kubenswrapper[4575]: I1004 04:58:48.422203 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2s2b6" event={"ID":"f8d694b1-ab26-49ff-97c6-4a26ef3a0142","Type":"ContainerDied","Data":"92ff953ac58e7b34212b6fb3f6650f4156d0273132432a20d45b37fd75280541"} Oct 04 04:58:49 crc kubenswrapper[4575]: I1004 04:58:49.831435 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:49 crc kubenswrapper[4575]: I1004 04:58:49.994830 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-combined-ca-bundle\") pod \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " Oct 04 04:58:49 crc kubenswrapper[4575]: I1004 04:58:49.994925 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-scripts\") pod \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " Oct 04 04:58:49 crc kubenswrapper[4575]: I1004 04:58:49.994983 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-config-data\") pod \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " Oct 04 04:58:49 crc kubenswrapper[4575]: I1004 04:58:49.995011 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp8bm\" (UniqueName: \"kubernetes.io/projected/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-kube-api-access-rp8bm\") pod \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\" (UID: \"f8d694b1-ab26-49ff-97c6-4a26ef3a0142\") " Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.001800 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-kube-api-access-rp8bm" (OuterVolumeSpecName: "kube-api-access-rp8bm") pod "f8d694b1-ab26-49ff-97c6-4a26ef3a0142" (UID: "f8d694b1-ab26-49ff-97c6-4a26ef3a0142"). InnerVolumeSpecName "kube-api-access-rp8bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.002030 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-scripts" (OuterVolumeSpecName: "scripts") pod "f8d694b1-ab26-49ff-97c6-4a26ef3a0142" (UID: "f8d694b1-ab26-49ff-97c6-4a26ef3a0142"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.026965 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-config-data" (OuterVolumeSpecName: "config-data") pod "f8d694b1-ab26-49ff-97c6-4a26ef3a0142" (UID: "f8d694b1-ab26-49ff-97c6-4a26ef3a0142"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.040887 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8d694b1-ab26-49ff-97c6-4a26ef3a0142" (UID: "f8d694b1-ab26-49ff-97c6-4a26ef3a0142"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.097112 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.097148 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp8bm\" (UniqueName: \"kubernetes.io/projected/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-kube-api-access-rp8bm\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.097164 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.097175 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d694b1-ab26-49ff-97c6-4a26ef3a0142-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.444501 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2s2b6" event={"ID":"f8d694b1-ab26-49ff-97c6-4a26ef3a0142","Type":"ContainerDied","Data":"6b5b7ba26cdac66f56744c60d1a9912b2ff09cc0ea4a7cf135700149bd496c2d"} Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.444554 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b5b7ba26cdac66f56744c60d1a9912b2ff09cc0ea4a7cf135700149bd496c2d" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.444634 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2s2b6" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.581645 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:58:50 crc kubenswrapper[4575]: E1004 04:58:50.583001 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d694b1-ab26-49ff-97c6-4a26ef3a0142" containerName="nova-cell0-conductor-db-sync" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.583094 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d694b1-ab26-49ff-97c6-4a26ef3a0142" containerName="nova-cell0-conductor-db-sync" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.583396 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d694b1-ab26-49ff-97c6-4a26ef3a0142" containerName="nova-cell0-conductor-db-sync" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.584125 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.586381 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-mz7g2" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.586775 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.597704 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.715383 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k5df\" (UniqueName: \"kubernetes.io/projected/1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2-kube-api-access-8k5df\") pod \"nova-cell0-conductor-0\" (UID: \"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.715911 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.715959 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.817130 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.817182 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.817236 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k5df\" (UniqueName: \"kubernetes.io/projected/1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2-kube-api-access-8k5df\") pod \"nova-cell0-conductor-0\" (UID: \"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.822160 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.823027 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.841006 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k5df\" (UniqueName: \"kubernetes.io/projected/1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2-kube-api-access-8k5df\") pod \"nova-cell0-conductor-0\" (UID: \"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:50 crc kubenswrapper[4575]: I1004 04:58:50.904257 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:51 crc kubenswrapper[4575]: I1004 04:58:51.377932 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:58:51 crc kubenswrapper[4575]: W1004 04:58:51.422186 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1db1ea6f_24fd_4c11_a2ba_dd47fc6edda2.slice/crio-ef8e9ed8f2b19b2f28ebf7a60f34d58ad31c91cd55934e5464b49e10d9a27f2a WatchSource:0}: Error finding container ef8e9ed8f2b19b2f28ebf7a60f34d58ad31c91cd55934e5464b49e10d9a27f2a: Status 404 returned error can't find the container with id ef8e9ed8f2b19b2f28ebf7a60f34d58ad31c91cd55934e5464b49e10d9a27f2a Oct 04 04:58:51 crc kubenswrapper[4575]: I1004 04:58:51.463746 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2","Type":"ContainerStarted","Data":"ef8e9ed8f2b19b2f28ebf7a60f34d58ad31c91cd55934e5464b49e10d9a27f2a"} Oct 04 04:58:52 crc kubenswrapper[4575]: I1004 04:58:52.473300 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2","Type":"ContainerStarted","Data":"18e3254ec25ddcc0cefda8ab861c3a92c21b8aba61c359679122b7c746ff3780"} Oct 04 04:58:52 crc kubenswrapper[4575]: I1004 04:58:52.476210 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 04:58:52 crc kubenswrapper[4575]: I1004 04:58:52.494810 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.494793511 podStartE2EDuration="2.494793511s" podCreationTimestamp="2025-10-04 04:58:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:58:52.493006189 +0000 UTC m=+1483.821565023" watchObservedRunningTime="2025-10-04 04:58:52.494793511 +0000 UTC m=+1483.823352325" Oct 04 04:58:52 crc kubenswrapper[4575]: I1004 04:58:52.565342 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.178:3000/\": dial tcp 10.217.0.178:3000: connect: connection refused" Oct 04 04:58:55 crc kubenswrapper[4575]: I1004 04:58:55.557712 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:58:55 crc kubenswrapper[4575]: I1004 04:58:55.558431 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:58:55 crc kubenswrapper[4575]: I1004 04:58:55.559134 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:58:55 crc kubenswrapper[4575]: I1004 04:58:55.714234 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:58:55 crc kubenswrapper[4575]: I1004 04:58:55.714302 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.523059 4575 generic.go:334] "Generic (PLEG): container finished" podID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerID="ba8f91ab188b7ef89bc5187a283e3172ed18fa55ed1e8f21e7bcf280a10203b7" exitCode=137 Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.523378 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerDied","Data":"ba8f91ab188b7ef89bc5187a283e3172ed18fa55ed1e8f21e7bcf280a10203b7"} Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.613859 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.746419 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-sg-core-conf-yaml\") pod \"c554dcf8-2717-4e95-809f-e61c318f29ca\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.746810 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-config-data\") pod \"c554dcf8-2717-4e95-809f-e61c318f29ca\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.747005 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-combined-ca-bundle\") pod \"c554dcf8-2717-4e95-809f-e61c318f29ca\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.747105 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbsx2\" (UniqueName: \"kubernetes.io/projected/c554dcf8-2717-4e95-809f-e61c318f29ca-kube-api-access-pbsx2\") pod \"c554dcf8-2717-4e95-809f-e61c318f29ca\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.747224 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-run-httpd\") pod \"c554dcf8-2717-4e95-809f-e61c318f29ca\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.747257 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-scripts\") pod \"c554dcf8-2717-4e95-809f-e61c318f29ca\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.747274 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-log-httpd\") pod \"c554dcf8-2717-4e95-809f-e61c318f29ca\" (UID: \"c554dcf8-2717-4e95-809f-e61c318f29ca\") " Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.748112 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c554dcf8-2717-4e95-809f-e61c318f29ca" (UID: "c554dcf8-2717-4e95-809f-e61c318f29ca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.748676 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c554dcf8-2717-4e95-809f-e61c318f29ca" (UID: "c554dcf8-2717-4e95-809f-e61c318f29ca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.767167 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c554dcf8-2717-4e95-809f-e61c318f29ca-kube-api-access-pbsx2" (OuterVolumeSpecName: "kube-api-access-pbsx2") pod "c554dcf8-2717-4e95-809f-e61c318f29ca" (UID: "c554dcf8-2717-4e95-809f-e61c318f29ca"). InnerVolumeSpecName "kube-api-access-pbsx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.767186 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-scripts" (OuterVolumeSpecName: "scripts") pod "c554dcf8-2717-4e95-809f-e61c318f29ca" (UID: "c554dcf8-2717-4e95-809f-e61c318f29ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.782704 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c554dcf8-2717-4e95-809f-e61c318f29ca" (UID: "c554dcf8-2717-4e95-809f-e61c318f29ca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.841095 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c554dcf8-2717-4e95-809f-e61c318f29ca" (UID: "c554dcf8-2717-4e95-809f-e61c318f29ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.849376 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbsx2\" (UniqueName: \"kubernetes.io/projected/c554dcf8-2717-4e95-809f-e61c318f29ca-kube-api-access-pbsx2\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.850036 4575 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.850137 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.850238 4575 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c554dcf8-2717-4e95-809f-e61c318f29ca-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.850310 4575 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.850409 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.852460 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-config-data" (OuterVolumeSpecName: "config-data") pod "c554dcf8-2717-4e95-809f-e61c318f29ca" (UID: "c554dcf8-2717-4e95-809f-e61c318f29ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:57 crc kubenswrapper[4575]: I1004 04:58:57.952897 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c554dcf8-2717-4e95-809f-e61c318f29ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.537648 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c554dcf8-2717-4e95-809f-e61c318f29ca","Type":"ContainerDied","Data":"3e1f474c96a167b189c034d7c9747fccd1a564628300c7eb3a1bf33204e00779"} Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.537695 4575 scope.go:117] "RemoveContainer" containerID="79df2b6454e6cb434758d00e83473834c649a5e7d058cd8e05d1c37befe32508" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.537948 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.572929 4575 scope.go:117] "RemoveContainer" containerID="f1d05f0a6d4c4ef80c6ba9cd1817181af964eab80fd5884781091596e8e65b1e" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.577460 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.593356 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.612615 4575 scope.go:117] "RemoveContainer" containerID="ba8f91ab188b7ef89bc5187a283e3172ed18fa55ed1e8f21e7bcf280a10203b7" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.619021 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:58 crc kubenswrapper[4575]: E1004 04:58:58.619476 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="proxy-httpd" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.619524 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="proxy-httpd" Oct 04 04:58:58 crc kubenswrapper[4575]: E1004 04:58:58.619548 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="sg-core" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.619555 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="sg-core" Oct 04 04:58:58 crc kubenswrapper[4575]: E1004 04:58:58.619597 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="ceilometer-notification-agent" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.619604 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="ceilometer-notification-agent" Oct 04 04:58:58 crc kubenswrapper[4575]: E1004 04:58:58.619619 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="ceilometer-central-agent" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.619625 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="ceilometer-central-agent" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.619796 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="ceilometer-central-agent" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.619815 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="ceilometer-notification-agent" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.619829 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="proxy-httpd" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.619838 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" containerName="sg-core" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.621997 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.626674 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.641511 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.641878 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.677028 4575 scope.go:117] "RemoveContainer" containerID="7ab01d41cf423474a2609ea04e1a8ba8cac932e91a8ede97711dc866cb3753e9" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.772028 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-run-httpd\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.772097 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.772138 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-log-httpd\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.772170 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-scripts\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.772205 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2kgn\" (UniqueName: \"kubernetes.io/projected/26a904de-6d26-41ce-a29b-7cd0dbe0f610-kube-api-access-c2kgn\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.772275 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-config-data\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.772297 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.873771 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-run-httpd\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.873821 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.873854 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-log-httpd\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.873881 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-scripts\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.874223 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-log-httpd\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.874286 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2kgn\" (UniqueName: \"kubernetes.io/projected/26a904de-6d26-41ce-a29b-7cd0dbe0f610-kube-api-access-c2kgn\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.874356 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-run-httpd\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.874726 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-config-data\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.874749 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.878824 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.895636 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.896084 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-config-data\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.896400 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-scripts\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.897004 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2kgn\" (UniqueName: \"kubernetes.io/projected/26a904de-6d26-41ce-a29b-7cd0dbe0f610-kube-api-access-c2kgn\") pod \"ceilometer-0\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " pod="openstack/ceilometer-0" Oct 04 04:58:58 crc kubenswrapper[4575]: I1004 04:58:58.961705 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.116312 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dtc62"] Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.118403 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.127701 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dtc62"] Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.287730 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-utilities\") pod \"redhat-operators-dtc62\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.288057 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-catalog-content\") pod \"redhat-operators-dtc62\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.288172 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj2b2\" (UniqueName: \"kubernetes.io/projected/c231de36-c828-4591-affb-cf345a773289-kube-api-access-nj2b2\") pod \"redhat-operators-dtc62\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.329189 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c554dcf8-2717-4e95-809f-e61c318f29ca" path="/var/lib/kubelet/pods/c554dcf8-2717-4e95-809f-e61c318f29ca/volumes" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.389210 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-utilities\") pod \"redhat-operators-dtc62\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.389261 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-catalog-content\") pod \"redhat-operators-dtc62\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.389349 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj2b2\" (UniqueName: \"kubernetes.io/projected/c231de36-c828-4591-affb-cf345a773289-kube-api-access-nj2b2\") pod \"redhat-operators-dtc62\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.389737 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-utilities\") pod \"redhat-operators-dtc62\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.389730 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-catalog-content\") pod \"redhat-operators-dtc62\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.415576 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj2b2\" (UniqueName: \"kubernetes.io/projected/c231de36-c828-4591-affb-cf345a773289-kube-api-access-nj2b2\") pod \"redhat-operators-dtc62\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.454381 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.473133 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.572026 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerStarted","Data":"c85851676b289c4e7eb2c406a43570e6af7dfe4c5e3a6de3ed9aa1b071e149f8"} Oct 04 04:58:59 crc kubenswrapper[4575]: I1004 04:58:59.950963 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dtc62"] Oct 04 04:58:59 crc kubenswrapper[4575]: W1004 04:58:59.951049 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc231de36_c828_4591_affb_cf345a773289.slice/crio-edd570faf2fef63d70c937024afd8e4766fd090ad402fd7899d06fa54120270d WatchSource:0}: Error finding container edd570faf2fef63d70c937024afd8e4766fd090ad402fd7899d06fa54120270d: Status 404 returned error can't find the container with id edd570faf2fef63d70c937024afd8e4766fd090ad402fd7899d06fa54120270d Oct 04 04:59:00 crc kubenswrapper[4575]: I1004 04:59:00.590458 4575 generic.go:334] "Generic (PLEG): container finished" podID="c231de36-c828-4591-affb-cf345a773289" containerID="406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea" exitCode=0 Oct 04 04:59:00 crc kubenswrapper[4575]: I1004 04:59:00.590849 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtc62" event={"ID":"c231de36-c828-4591-affb-cf345a773289","Type":"ContainerDied","Data":"406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea"} Oct 04 04:59:00 crc kubenswrapper[4575]: I1004 04:59:00.590883 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtc62" event={"ID":"c231de36-c828-4591-affb-cf345a773289","Type":"ContainerStarted","Data":"edd570faf2fef63d70c937024afd8e4766fd090ad402fd7899d06fa54120270d"} Oct 04 04:59:00 crc kubenswrapper[4575]: I1004 04:59:00.595758 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerStarted","Data":"7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0"} Oct 04 04:59:00 crc kubenswrapper[4575]: I1004 04:59:00.932402 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 04:59:01 crc kubenswrapper[4575]: I1004 04:59:01.630193 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerStarted","Data":"a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9"} Oct 04 04:59:01 crc kubenswrapper[4575]: I1004 04:59:01.933327 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-pldbk"] Oct 04 04:59:01 crc kubenswrapper[4575]: I1004 04:59:01.936524 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:01 crc kubenswrapper[4575]: I1004 04:59:01.944337 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 04 04:59:01 crc kubenswrapper[4575]: I1004 04:59:01.944560 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 04 04:59:01 crc kubenswrapper[4575]: I1004 04:59:01.959685 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-pldbk"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.052780 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-scripts\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.052917 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-config-data\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.053005 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.053044 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mp2g\" (UniqueName: \"kubernetes.io/projected/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-kube-api-access-5mp2g\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.156279 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-scripts\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.156363 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-config-data\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.156435 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.156478 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mp2g\" (UniqueName: \"kubernetes.io/projected/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-kube-api-access-5mp2g\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.173780 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-scripts\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.178444 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-config-data\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.185179 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.218502 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.221917 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.222807 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mp2g\" (UniqueName: \"kubernetes.io/projected/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-kube-api-access-5mp2g\") pod \"nova-cell0-cell-mapping-pldbk\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.231307 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.249903 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.263623 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.370154 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2fgr\" (UniqueName: \"kubernetes.io/projected/75082b13-cf45-44ec-a2a5-f7210ba72289-kube-api-access-x2fgr\") pod \"nova-scheduler-0\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.370220 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-config-data\") pod \"nova-scheduler-0\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.370260 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.389160 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.400773 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.405336 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.429360 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.430904 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.441331 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.474790 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2fgr\" (UniqueName: \"kubernetes.io/projected/75082b13-cf45-44ec-a2a5-f7210ba72289-kube-api-access-x2fgr\") pod \"nova-scheduler-0\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.474872 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-config-data\") pod \"nova-scheduler-0\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.474936 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.500708 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.505376 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.518345 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-config-data\") pod \"nova-scheduler-0\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.522384 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2fgr\" (UniqueName: \"kubernetes.io/projected/75082b13-cf45-44ec-a2a5-f7210ba72289-kube-api-access-x2fgr\") pod \"nova-scheduler-0\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.529580 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.576423 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18302821-ac2c-4ca2-9f33-86ba861a9e0a-logs\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.576500 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-config-data\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.576551 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpk84\" (UniqueName: \"kubernetes.io/projected/18302821-ac2c-4ca2-9f33-86ba861a9e0a-kube-api-access-cpk84\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.576772 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.576808 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.576876 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdpwq\" (UniqueName: \"kubernetes.io/projected/915e37b0-2385-42b9-b9fe-a4b1612e00d6-kube-api-access-fdpwq\") pod \"nova-cell1-novncproxy-0\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.576945 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.604674 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.629747 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.631948 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.647560 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.676381 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vn9rj"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.680573 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.682250 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.682381 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18302821-ac2c-4ca2-9f33-86ba861a9e0a-logs\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.682467 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-config-data\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.682543 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpk84\" (UniqueName: \"kubernetes.io/projected/18302821-ac2c-4ca2-9f33-86ba861a9e0a-kube-api-access-cpk84\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.682622 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.682660 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.682770 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdpwq\" (UniqueName: \"kubernetes.io/projected/915e37b0-2385-42b9-b9fe-a4b1612e00d6-kube-api-access-fdpwq\") pod \"nova-cell1-novncproxy-0\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.684369 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18302821-ac2c-4ca2-9f33-86ba861a9e0a-logs\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.700959 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.721225 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.721291 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-config-data\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.753580 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.754216 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdpwq\" (UniqueName: \"kubernetes.io/projected/915e37b0-2385-42b9-b9fe-a4b1612e00d6-kube-api-access-fdpwq\") pod \"nova-cell1-novncproxy-0\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.754528 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.760065 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerStarted","Data":"e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1"} Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.796448 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtc62" event={"ID":"c231de36-c828-4591-affb-cf345a773289","Type":"ContainerStarted","Data":"72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a"} Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.800637 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.800742 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.801016 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.801076 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-config\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.801100 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.801272 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-config-data\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.801303 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-svc\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.801373 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd55d\" (UniqueName: \"kubernetes.io/projected/8b477b40-b8de-400b-9caa-e38540b59c47-kube-api-access-jd55d\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.801412 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bjxg\" (UniqueName: \"kubernetes.io/projected/854288d3-ecc2-482f-a986-f27cc7672efa-kube-api-access-4bjxg\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.801447 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b477b40-b8de-400b-9caa-e38540b59c47-logs\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.811766 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpk84\" (UniqueName: \"kubernetes.io/projected/18302821-ac2c-4ca2-9f33-86ba861a9e0a-kube-api-access-cpk84\") pod \"nova-metadata-0\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " pod="openstack/nova-metadata-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.888975 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vn9rj"] Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906023 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906098 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-config\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906126 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906199 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-config-data\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906218 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-svc\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906244 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd55d\" (UniqueName: \"kubernetes.io/projected/8b477b40-b8de-400b-9caa-e38540b59c47-kube-api-access-jd55d\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906268 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bjxg\" (UniqueName: \"kubernetes.io/projected/854288d3-ecc2-482f-a986-f27cc7672efa-kube-api-access-4bjxg\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906288 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b477b40-b8de-400b-9caa-e38540b59c47-logs\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906335 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.906363 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.907990 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b477b40-b8de-400b-9caa-e38540b59c47-logs\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.907990 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-svc\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.908556 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-config\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.911508 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.912314 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.918170 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.929166 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-config-data\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.935171 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bjxg\" (UniqueName: \"kubernetes.io/projected/854288d3-ecc2-482f-a986-f27cc7672efa-kube-api-access-4bjxg\") pod \"dnsmasq-dns-865f5d856f-vn9rj\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.935978 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.947800 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd55d\" (UniqueName: \"kubernetes.io/projected/8b477b40-b8de-400b-9caa-e38540b59c47-kube-api-access-jd55d\") pod \"nova-api-0\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " pod="openstack/nova-api-0" Oct 04 04:59:02 crc kubenswrapper[4575]: I1004 04:59:02.955741 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:59:03 crc kubenswrapper[4575]: I1004 04:59:03.059152 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:03 crc kubenswrapper[4575]: I1004 04:59:03.176394 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:59:03 crc kubenswrapper[4575]: I1004 04:59:03.221367 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:03 crc kubenswrapper[4575]: I1004 04:59:03.641078 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-pldbk"] Oct 04 04:59:03 crc kubenswrapper[4575]: I1004 04:59:03.852274 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerStarted","Data":"1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876"} Oct 04 04:59:03 crc kubenswrapper[4575]: I1004 04:59:03.853342 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:59:03 crc kubenswrapper[4575]: I1004 04:59:03.862404 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pldbk" event={"ID":"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8","Type":"ContainerStarted","Data":"fdceed21224c6238a49ea06189e8e7db718ecb081c0a5f3105203a7ad80b495f"} Oct 04 04:59:03 crc kubenswrapper[4575]: I1004 04:59:03.888950 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:59:03 crc kubenswrapper[4575]: I1004 04:59:03.896526 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.527451062 podStartE2EDuration="5.896505667s" podCreationTimestamp="2025-10-04 04:58:58 +0000 UTC" firstStartedPulling="2025-10-04 04:58:59.483146098 +0000 UTC m=+1490.811704912" lastFinishedPulling="2025-10-04 04:59:02.852200703 +0000 UTC m=+1494.180759517" observedRunningTime="2025-10-04 04:59:03.891853423 +0000 UTC m=+1495.220412247" watchObservedRunningTime="2025-10-04 04:59:03.896505667 +0000 UTC m=+1495.225064481" Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.013162 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.067106 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.370423 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.391932 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vn9rj"] Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.881984 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8b477b40-b8de-400b-9caa-e38540b59c47","Type":"ContainerStarted","Data":"50e8c4ea93e8cb6db66947a2356f61f4d37da1467e80737faa491029afa0e025"} Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.884686 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"75082b13-cf45-44ec-a2a5-f7210ba72289","Type":"ContainerStarted","Data":"6ff5c863d1d316972c2d519ad205f6dab0057562fc129e27dcc442e8677a3141"} Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.895849 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pldbk" event={"ID":"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8","Type":"ContainerStarted","Data":"e113320cd2896101d1b308ab535199ca02eabed2946ee283bc62645c482dcc72"} Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.902178 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"915e37b0-2385-42b9-b9fe-a4b1612e00d6","Type":"ContainerStarted","Data":"95ea5541ca348f0d49fd94be947f9e0f6d29407a165418275c960be43871d5dc"} Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.934107 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-pldbk" podStartSLOduration=3.934085198 podStartE2EDuration="3.934085198s" podCreationTimestamp="2025-10-04 04:59:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:04.92966042 +0000 UTC m=+1496.258219234" watchObservedRunningTime="2025-10-04 04:59:04.934085198 +0000 UTC m=+1496.262644012" Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.939413 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18302821-ac2c-4ca2-9f33-86ba861a9e0a","Type":"ContainerStarted","Data":"ab4fc1cfd385d3f21c1a6ba0cf2bb39099c474db7650db1e9f275ad4e10b6f57"} Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.950107 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" event={"ID":"854288d3-ecc2-482f-a986-f27cc7672efa","Type":"ContainerStarted","Data":"2d538297cbfbed345effb9766b528351743d21527348d089232432a7faad8a04"} Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.992504 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-znqtq"] Oct 04 04:59:04 crc kubenswrapper[4575]: I1004 04:59:04.993780 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.002060 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.002561 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.025142 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-znqtq"] Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.090752 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.092080 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-scripts\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.092156 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-config-data\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.092181 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbcqj\" (UniqueName: \"kubernetes.io/projected/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-kube-api-access-tbcqj\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.193486 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-scripts\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.193538 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-config-data\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.193557 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbcqj\" (UniqueName: \"kubernetes.io/projected/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-kube-api-access-tbcqj\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.193712 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.213880 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-config-data\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.218148 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-scripts\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.218467 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.227243 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbcqj\" (UniqueName: \"kubernetes.io/projected/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-kube-api-access-tbcqj\") pod \"nova-cell1-conductor-db-sync-znqtq\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.347128 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.562078 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.720960 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.990856 4575 generic.go:334] "Generic (PLEG): container finished" podID="854288d3-ecc2-482f-a986-f27cc7672efa" containerID="2c5d998293c916864824f68c020bf1c52dafb1c55451382aabf4716071c249e5" exitCode=0 Oct 04 04:59:05 crc kubenswrapper[4575]: I1004 04:59:05.992550 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" event={"ID":"854288d3-ecc2-482f-a986-f27cc7672efa","Type":"ContainerDied","Data":"2c5d998293c916864824f68c020bf1c52dafb1c55451382aabf4716071c249e5"} Oct 04 04:59:06 crc kubenswrapper[4575]: I1004 04:59:05.997193 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-znqtq"] Oct 04 04:59:06 crc kubenswrapper[4575]: I1004 04:59:06.955118 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:59:07 crc kubenswrapper[4575]: I1004 04:59:07.007678 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:59:07 crc kubenswrapper[4575]: I1004 04:59:07.035485 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" event={"ID":"854288d3-ecc2-482f-a986-f27cc7672efa","Type":"ContainerStarted","Data":"f4bb2f72787716c30050fdbe6afcc0302e5cd9b363fd6387254b80f9896ff6bd"} Oct 04 04:59:07 crc kubenswrapper[4575]: I1004 04:59:07.035563 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:07 crc kubenswrapper[4575]: I1004 04:59:07.040041 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-znqtq" event={"ID":"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74","Type":"ContainerStarted","Data":"5ba8d76ad793e81069c3f3e732151ec619db7ced0cd6d6118b6effe957a41c5e"} Oct 04 04:59:07 crc kubenswrapper[4575]: I1004 04:59:07.040090 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-znqtq" event={"ID":"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74","Type":"ContainerStarted","Data":"f5734b42efd5abbc026d4ece064b6dbde6f28cddc6263f0765661609eb85dd16"} Oct 04 04:59:07 crc kubenswrapper[4575]: I1004 04:59:07.060930 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" podStartSLOduration=5.060911309 podStartE2EDuration="5.060911309s" podCreationTimestamp="2025-10-04 04:59:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:07.059476767 +0000 UTC m=+1498.388035601" watchObservedRunningTime="2025-10-04 04:59:07.060911309 +0000 UTC m=+1498.389470123" Oct 04 04:59:07 crc kubenswrapper[4575]: I1004 04:59:07.082618 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-znqtq" podStartSLOduration=3.082193062 podStartE2EDuration="3.082193062s" podCreationTimestamp="2025-10-04 04:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:07.077226459 +0000 UTC m=+1498.405785273" watchObservedRunningTime="2025-10-04 04:59:07.082193062 +0000 UTC m=+1498.410751896" Oct 04 04:59:09 crc kubenswrapper[4575]: I1004 04:59:09.098203 4575 generic.go:334] "Generic (PLEG): container finished" podID="c231de36-c828-4591-affb-cf345a773289" containerID="72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a" exitCode=0 Oct 04 04:59:09 crc kubenswrapper[4575]: I1004 04:59:09.098397 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtc62" event={"ID":"c231de36-c828-4591-affb-cf345a773289","Type":"ContainerDied","Data":"72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a"} Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.179821 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8b477b40-b8de-400b-9caa-e38540b59c47","Type":"ContainerStarted","Data":"f61d6d584e189139f2e51e9ce576f8d5d56641866bfbf7a2dd52539b5a2a704a"} Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.201137 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"75082b13-cf45-44ec-a2a5-f7210ba72289","Type":"ContainerStarted","Data":"9b8408fdb07e2e8da5bb7a92bc9bfa21a078482b939c9cf3d3ce6d013498d2c7"} Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.202855 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"915e37b0-2385-42b9-b9fe-a4b1612e00d6","Type":"ContainerStarted","Data":"3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb"} Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.202988 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="915e37b0-2385-42b9-b9fe-a4b1612e00d6" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb" gracePeriod=30 Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.217631 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtc62" event={"ID":"c231de36-c828-4591-affb-cf345a773289","Type":"ContainerStarted","Data":"3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459"} Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.243924 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.906552774 podStartE2EDuration="10.243907129s" podCreationTimestamp="2025-10-04 04:59:02 +0000 UTC" firstStartedPulling="2025-10-04 04:59:03.909952405 +0000 UTC m=+1495.238511219" lastFinishedPulling="2025-10-04 04:59:11.24730676 +0000 UTC m=+1502.575865574" observedRunningTime="2025-10-04 04:59:12.238105532 +0000 UTC m=+1503.566664346" watchObservedRunningTime="2025-10-04 04:59:12.243907129 +0000 UTC m=+1503.572465943" Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.269965 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.032810195 podStartE2EDuration="10.269944399s" podCreationTimestamp="2025-10-04 04:59:02 +0000 UTC" firstStartedPulling="2025-10-04 04:59:04.027375263 +0000 UTC m=+1495.355934077" lastFinishedPulling="2025-10-04 04:59:11.264509467 +0000 UTC m=+1502.593068281" observedRunningTime="2025-10-04 04:59:12.266033506 +0000 UTC m=+1503.594592320" watchObservedRunningTime="2025-10-04 04:59:12.269944399 +0000 UTC m=+1503.598503213" Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.341286 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dtc62" podStartSLOduration=2.670243797 podStartE2EDuration="13.341267926s" podCreationTimestamp="2025-10-04 04:58:59 +0000 UTC" firstStartedPulling="2025-10-04 04:59:00.593256821 +0000 UTC m=+1491.921815645" lastFinishedPulling="2025-10-04 04:59:11.26428097 +0000 UTC m=+1502.592839774" observedRunningTime="2025-10-04 04:59:12.327491479 +0000 UTC m=+1503.656050323" watchObservedRunningTime="2025-10-04 04:59:12.341267926 +0000 UTC m=+1503.669826750" Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.609568 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.609628 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.650122 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.975486 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4ml2t"] Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.987869 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:12 crc kubenswrapper[4575]: I1004 04:59:12.991993 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4ml2t"] Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.060328 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.078702 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-utilities\") pod \"community-operators-4ml2t\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.078809 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-catalog-content\") pod \"community-operators-4ml2t\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.078862 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4dgp\" (UniqueName: \"kubernetes.io/projected/8da33bfd-1332-4c02-b37a-3605982815e8-kube-api-access-f4dgp\") pod \"community-operators-4ml2t\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.180619 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-utilities\") pod \"community-operators-4ml2t\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.180758 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-catalog-content\") pod \"community-operators-4ml2t\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.180828 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4dgp\" (UniqueName: \"kubernetes.io/projected/8da33bfd-1332-4c02-b37a-3605982815e8-kube-api-access-f4dgp\") pod \"community-operators-4ml2t\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.181184 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-utilities\") pod \"community-operators-4ml2t\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.181574 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-catalog-content\") pod \"community-operators-4ml2t\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.197088 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4dgp\" (UniqueName: \"kubernetes.io/projected/8da33bfd-1332-4c02-b37a-3605982815e8-kube-api-access-f4dgp\") pod \"community-operators-4ml2t\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.224675 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.228475 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18302821-ac2c-4ca2-9f33-86ba861a9e0a","Type":"ContainerStarted","Data":"96385a08097f46b617b586399b842fecc52bc222eec389ab74b37fba5f2df3e2"} Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.228517 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18302821-ac2c-4ca2-9f33-86ba861a9e0a","Type":"ContainerStarted","Data":"3910011c01da877db2d770d227425f7450dff968195a5879ab0011eb13e09eda"} Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.228571 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerName="nova-metadata-metadata" containerID="cri-o://96385a08097f46b617b586399b842fecc52bc222eec389ab74b37fba5f2df3e2" gracePeriod=30 Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.228780 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerName="nova-metadata-log" containerID="cri-o://3910011c01da877db2d770d227425f7450dff968195a5879ab0011eb13e09eda" gracePeriod=30 Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.241793 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8b477b40-b8de-400b-9caa-e38540b59c47","Type":"ContainerStarted","Data":"865c212f049bf39aec2d2c254afacf51d842fae1342f99c2bf42dbfc9f22dbb8"} Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.289772 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.408362064 podStartE2EDuration="11.289658384s" podCreationTimestamp="2025-10-04 04:59:02 +0000 UTC" firstStartedPulling="2025-10-04 04:59:04.388420217 +0000 UTC m=+1495.716979021" lastFinishedPulling="2025-10-04 04:59:11.269716527 +0000 UTC m=+1502.598275341" observedRunningTime="2025-10-04 04:59:13.279310206 +0000 UTC m=+1504.607869040" watchObservedRunningTime="2025-10-04 04:59:13.289658384 +0000 UTC m=+1504.618217198" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.315910 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.346340 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.355649 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-fnb6x"] Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.356072 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" podUID="b7b66914-2f2d-4068-ae3e-2b742543d07d" containerName="dnsmasq-dns" containerID="cri-o://96642340d0958fc56f29cce3f95246afb1a669c35882f75cff8dfe66e7b655f2" gracePeriod=10 Oct 04 04:59:13 crc kubenswrapper[4575]: I1004 04:59:13.376018 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.137187403 podStartE2EDuration="11.375996255s" podCreationTimestamp="2025-10-04 04:59:02 +0000 UTC" firstStartedPulling="2025-10-04 04:59:04.024394537 +0000 UTC m=+1495.352953341" lastFinishedPulling="2025-10-04 04:59:11.263203379 +0000 UTC m=+1502.591762193" observedRunningTime="2025-10-04 04:59:13.32625839 +0000 UTC m=+1504.654817204" watchObservedRunningTime="2025-10-04 04:59:13.375996255 +0000 UTC m=+1504.704555069" Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.058391 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4ml2t"] Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.272962 4575 generic.go:334] "Generic (PLEG): container finished" podID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerID="3910011c01da877db2d770d227425f7450dff968195a5879ab0011eb13e09eda" exitCode=143 Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.275923 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18302821-ac2c-4ca2-9f33-86ba861a9e0a","Type":"ContainerDied","Data":"3910011c01da877db2d770d227425f7450dff968195a5879ab0011eb13e09eda"} Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.292649 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4ml2t" event={"ID":"8da33bfd-1332-4c02-b37a-3605982815e8","Type":"ContainerStarted","Data":"56c16780e51bdb2d665313fb5e573d2c575c5bb1d4a34699b5e8a262fbe51114"} Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.305056 4575 generic.go:334] "Generic (PLEG): container finished" podID="b7b66914-2f2d-4068-ae3e-2b742543d07d" containerID="96642340d0958fc56f29cce3f95246afb1a669c35882f75cff8dfe66e7b655f2" exitCode=0 Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.305357 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" event={"ID":"b7b66914-2f2d-4068-ae3e-2b742543d07d","Type":"ContainerDied","Data":"96642340d0958fc56f29cce3f95246afb1a669c35882f75cff8dfe66e7b655f2"} Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.772448 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.949707 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np8nn\" (UniqueName: \"kubernetes.io/projected/b7b66914-2f2d-4068-ae3e-2b742543d07d-kube-api-access-np8nn\") pod \"b7b66914-2f2d-4068-ae3e-2b742543d07d\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.949790 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-config\") pod \"b7b66914-2f2d-4068-ae3e-2b742543d07d\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.950047 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-sb\") pod \"b7b66914-2f2d-4068-ae3e-2b742543d07d\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.950252 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-swift-storage-0\") pod \"b7b66914-2f2d-4068-ae3e-2b742543d07d\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.950356 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-nb\") pod \"b7b66914-2f2d-4068-ae3e-2b742543d07d\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.950388 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-svc\") pod \"b7b66914-2f2d-4068-ae3e-2b742543d07d\" (UID: \"b7b66914-2f2d-4068-ae3e-2b742543d07d\") " Oct 04 04:59:14 crc kubenswrapper[4575]: I1004 04:59:14.966265 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7b66914-2f2d-4068-ae3e-2b742543d07d-kube-api-access-np8nn" (OuterVolumeSpecName: "kube-api-access-np8nn") pod "b7b66914-2f2d-4068-ae3e-2b742543d07d" (UID: "b7b66914-2f2d-4068-ae3e-2b742543d07d"). InnerVolumeSpecName "kube-api-access-np8nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.054836 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np8nn\" (UniqueName: \"kubernetes.io/projected/b7b66914-2f2d-4068-ae3e-2b742543d07d-kube-api-access-np8nn\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.067820 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b7b66914-2f2d-4068-ae3e-2b742543d07d" (UID: "b7b66914-2f2d-4068-ae3e-2b742543d07d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.082538 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b7b66914-2f2d-4068-ae3e-2b742543d07d" (UID: "b7b66914-2f2d-4068-ae3e-2b742543d07d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.087387 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b7b66914-2f2d-4068-ae3e-2b742543d07d" (UID: "b7b66914-2f2d-4068-ae3e-2b742543d07d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.130200 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b7b66914-2f2d-4068-ae3e-2b742543d07d" (UID: "b7b66914-2f2d-4068-ae3e-2b742543d07d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.132509 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-config" (OuterVolumeSpecName: "config") pod "b7b66914-2f2d-4068-ae3e-2b742543d07d" (UID: "b7b66914-2f2d-4068-ae3e-2b742543d07d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.157144 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.157187 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.157196 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.157207 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.157218 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7b66914-2f2d-4068-ae3e-2b742543d07d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.327544 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.328114 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-fnb6x" event={"ID":"b7b66914-2f2d-4068-ae3e-2b742543d07d","Type":"ContainerDied","Data":"21a52c76ebac452ee83b55b1ec49d9bd3148f6e94c41a7777add746520af5368"} Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.336028 4575 scope.go:117] "RemoveContainer" containerID="96642340d0958fc56f29cce3f95246afb1a669c35882f75cff8dfe66e7b655f2" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.341007 4575 generic.go:334] "Generic (PLEG): container finished" podID="8da33bfd-1332-4c02-b37a-3605982815e8" containerID="c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244" exitCode=0 Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.342520 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4ml2t" event={"ID":"8da33bfd-1332-4c02-b37a-3605982815e8","Type":"ContainerDied","Data":"c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244"} Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.415118 4575 scope.go:117] "RemoveContainer" containerID="9be5659b5c55612b221fe9399b00f592fa94fba6eeb0ad41b3f9a5cad387870d" Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.488696 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-fnb6x"] Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.499121 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-fnb6x"] Oct 04 04:59:15 crc kubenswrapper[4575]: I1004 04:59:15.714128 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:59:16 crc kubenswrapper[4575]: I1004 04:59:16.350570 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4ml2t" event={"ID":"8da33bfd-1332-4c02-b37a-3605982815e8","Type":"ContainerStarted","Data":"42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af"} Oct 04 04:59:16 crc kubenswrapper[4575]: I1004 04:59:16.962826 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qj7qr"] Oct 04 04:59:16 crc kubenswrapper[4575]: E1004 04:59:16.963342 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b66914-2f2d-4068-ae3e-2b742543d07d" containerName="init" Oct 04 04:59:16 crc kubenswrapper[4575]: I1004 04:59:16.963361 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b66914-2f2d-4068-ae3e-2b742543d07d" containerName="init" Oct 04 04:59:16 crc kubenswrapper[4575]: E1004 04:59:16.963389 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b66914-2f2d-4068-ae3e-2b742543d07d" containerName="dnsmasq-dns" Oct 04 04:59:16 crc kubenswrapper[4575]: I1004 04:59:16.963397 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b66914-2f2d-4068-ae3e-2b742543d07d" containerName="dnsmasq-dns" Oct 04 04:59:16 crc kubenswrapper[4575]: I1004 04:59:16.963647 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7b66914-2f2d-4068-ae3e-2b742543d07d" containerName="dnsmasq-dns" Oct 04 04:59:16 crc kubenswrapper[4575]: I1004 04:59:16.965192 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:16 crc kubenswrapper[4575]: I1004 04:59:16.974778 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qj7qr"] Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.117621 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74bsl\" (UniqueName: \"kubernetes.io/projected/2b91f552-6eb5-479d-9a79-07a356d33ab0-kube-api-access-74bsl\") pod \"certified-operators-qj7qr\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.117732 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-utilities\") pod \"certified-operators-qj7qr\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.117763 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-catalog-content\") pod \"certified-operators-qj7qr\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.219502 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74bsl\" (UniqueName: \"kubernetes.io/projected/2b91f552-6eb5-479d-9a79-07a356d33ab0-kube-api-access-74bsl\") pod \"certified-operators-qj7qr\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.220057 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-utilities\") pod \"certified-operators-qj7qr\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.220193 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-catalog-content\") pod \"certified-operators-qj7qr\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.220506 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-utilities\") pod \"certified-operators-qj7qr\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.220915 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-catalog-content\") pod \"certified-operators-qj7qr\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.252946 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74bsl\" (UniqueName: \"kubernetes.io/projected/2b91f552-6eb5-479d-9a79-07a356d33ab0-kube-api-access-74bsl\") pod \"certified-operators-qj7qr\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.287267 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.326097 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7b66914-2f2d-4068-ae3e-2b742543d07d" path="/var/lib/kubelet/pods/b7b66914-2f2d-4068-ae3e-2b742543d07d/volumes" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.957089 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.957421 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:59:17 crc kubenswrapper[4575]: W1004 04:59:17.962872 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b91f552_6eb5_479d_9a79_07a356d33ab0.slice/crio-67007a30e75dbb01dfaf2954e83a968bd59ba71005fd6aac4acdc7055fab3898 WatchSource:0}: Error finding container 67007a30e75dbb01dfaf2954e83a968bd59ba71005fd6aac4acdc7055fab3898: Status 404 returned error can't find the container with id 67007a30e75dbb01dfaf2954e83a968bd59ba71005fd6aac4acdc7055fab3898 Oct 04 04:59:17 crc kubenswrapper[4575]: I1004 04:59:17.966479 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qj7qr"] Oct 04 04:59:18 crc kubenswrapper[4575]: I1004 04:59:18.380169 4575 generic.go:334] "Generic (PLEG): container finished" podID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerID="a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a" exitCode=0 Oct 04 04:59:18 crc kubenswrapper[4575]: I1004 04:59:18.380210 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj7qr" event={"ID":"2b91f552-6eb5-479d-9a79-07a356d33ab0","Type":"ContainerDied","Data":"a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a"} Oct 04 04:59:18 crc kubenswrapper[4575]: I1004 04:59:18.380235 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj7qr" event={"ID":"2b91f552-6eb5-479d-9a79-07a356d33ab0","Type":"ContainerStarted","Data":"67007a30e75dbb01dfaf2954e83a968bd59ba71005fd6aac4acdc7055fab3898"} Oct 04 04:59:19 crc kubenswrapper[4575]: I1004 04:59:19.417764 4575 generic.go:334] "Generic (PLEG): container finished" podID="8da33bfd-1332-4c02-b37a-3605982815e8" containerID="42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af" exitCode=0 Oct 04 04:59:19 crc kubenswrapper[4575]: I1004 04:59:19.418314 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4ml2t" event={"ID":"8da33bfd-1332-4c02-b37a-3605982815e8","Type":"ContainerDied","Data":"42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af"} Oct 04 04:59:19 crc kubenswrapper[4575]: I1004 04:59:19.455164 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:59:19 crc kubenswrapper[4575]: I1004 04:59:19.455508 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 04:59:20 crc kubenswrapper[4575]: I1004 04:59:20.429761 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj7qr" event={"ID":"2b91f552-6eb5-479d-9a79-07a356d33ab0","Type":"ContainerStarted","Data":"3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989"} Oct 04 04:59:20 crc kubenswrapper[4575]: I1004 04:59:20.431203 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4ml2t" event={"ID":"8da33bfd-1332-4c02-b37a-3605982815e8","Type":"ContainerStarted","Data":"eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028"} Oct 04 04:59:20 crc kubenswrapper[4575]: I1004 04:59:20.488975 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4ml2t" podStartSLOduration=3.930156743 podStartE2EDuration="8.488955917s" podCreationTimestamp="2025-10-04 04:59:12 +0000 UTC" firstStartedPulling="2025-10-04 04:59:15.346162757 +0000 UTC m=+1506.674721571" lastFinishedPulling="2025-10-04 04:59:19.904961931 +0000 UTC m=+1511.233520745" observedRunningTime="2025-10-04 04:59:20.483229652 +0000 UTC m=+1511.811788496" watchObservedRunningTime="2025-10-04 04:59:20.488955917 +0000 UTC m=+1511.817514731" Oct 04 04:59:20 crc kubenswrapper[4575]: I1004 04:59:20.564831 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:59:20 crc kubenswrapper[4575]: I1004 04:59:20.564919 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:59:20 crc kubenswrapper[4575]: I1004 04:59:20.565834 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"376a8e8470b9959c733967cacb7a6d6b2b157401d6330173cf7a8725128fbf8f"} pod="openstack/horizon-b86b879b4-tlf42" containerMessage="Container horizon failed startup probe, will be restarted" Oct 04 04:59:20 crc kubenswrapper[4575]: I1004 04:59:20.565914 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" containerID="cri-o://376a8e8470b9959c733967cacb7a6d6b2b157401d6330173cf7a8725128fbf8f" gracePeriod=30 Oct 04 04:59:20 crc kubenswrapper[4575]: I1004 04:59:20.631352 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtc62" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:20 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:20 crc kubenswrapper[4575]: > Oct 04 04:59:21 crc kubenswrapper[4575]: I1004 04:59:21.443283 4575 generic.go:334] "Generic (PLEG): container finished" podID="851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8" containerID="e113320cd2896101d1b308ab535199ca02eabed2946ee283bc62645c482dcc72" exitCode=0 Oct 04 04:59:21 crc kubenswrapper[4575]: I1004 04:59:21.443423 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pldbk" event={"ID":"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8","Type":"ContainerDied","Data":"e113320cd2896101d1b308ab535199ca02eabed2946ee283bc62645c482dcc72"} Oct 04 04:59:21 crc kubenswrapper[4575]: I1004 04:59:21.448516 4575 generic.go:334] "Generic (PLEG): container finished" podID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerID="3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989" exitCode=0 Oct 04 04:59:21 crc kubenswrapper[4575]: I1004 04:59:21.448557 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj7qr" event={"ID":"2b91f552-6eb5-479d-9a79-07a356d33ab0","Type":"ContainerDied","Data":"3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989"} Oct 04 04:59:22 crc kubenswrapper[4575]: I1004 04:59:22.472792 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj7qr" event={"ID":"2b91f552-6eb5-479d-9a79-07a356d33ab0","Type":"ContainerStarted","Data":"a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c"} Oct 04 04:59:22 crc kubenswrapper[4575]: I1004 04:59:22.509831 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qj7qr" podStartSLOduration=3.053732517 podStartE2EDuration="6.509813672s" podCreationTimestamp="2025-10-04 04:59:16 +0000 UTC" firstStartedPulling="2025-10-04 04:59:18.381948958 +0000 UTC m=+1509.710507772" lastFinishedPulling="2025-10-04 04:59:21.838030123 +0000 UTC m=+1513.166588927" observedRunningTime="2025-10-04 04:59:22.496037554 +0000 UTC m=+1513.824596378" watchObservedRunningTime="2025-10-04 04:59:22.509813672 +0000 UTC m=+1513.838372486" Oct 04 04:59:22 crc kubenswrapper[4575]: I1004 04:59:22.938871 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.090656 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-scripts\") pod \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.090703 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-combined-ca-bundle\") pod \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.090762 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-config-data\") pod \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.090842 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mp2g\" (UniqueName: \"kubernetes.io/projected/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-kube-api-access-5mp2g\") pod \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\" (UID: \"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8\") " Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.100145 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-scripts" (OuterVolumeSpecName: "scripts") pod "851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8" (UID: "851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.100727 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-kube-api-access-5mp2g" (OuterVolumeSpecName: "kube-api-access-5mp2g") pod "851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8" (UID: "851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8"). InnerVolumeSpecName "kube-api-access-5mp2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.136111 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-config-data" (OuterVolumeSpecName: "config-data") pod "851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8" (UID: "851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.137768 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8" (UID: "851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.177873 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.177939 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.192779 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.192815 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.192830 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.192843 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mp2g\" (UniqueName: \"kubernetes.io/projected/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8-kube-api-access-5mp2g\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.321758 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.321815 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.482563 4575 generic.go:334] "Generic (PLEG): container finished" podID="6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74" containerID="5ba8d76ad793e81069c3f3e732151ec619db7ced0cd6d6118b6effe957a41c5e" exitCode=0 Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.482745 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-znqtq" event={"ID":"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74","Type":"ContainerDied","Data":"5ba8d76ad793e81069c3f3e732151ec619db7ced0cd6d6118b6effe957a41c5e"} Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.485320 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pldbk" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.485818 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pldbk" event={"ID":"851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8","Type":"ContainerDied","Data":"fdceed21224c6238a49ea06189e8e7db718ecb081c0a5f3105203a7ad80b495f"} Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.485843 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdceed21224c6238a49ea06189e8e7db718ecb081c0a5f3105203a7ad80b495f" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.663597 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.664087 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-log" containerID="cri-o://f61d6d584e189139f2e51e9ce576f8d5d56641866bfbf7a2dd52539b5a2a704a" gracePeriod=30 Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.664176 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-api" containerID="cri-o://865c212f049bf39aec2d2c254afacf51d842fae1342f99c2bf42dbfc9f22dbb8" gracePeriod=30 Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.679679 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.679917 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="75082b13-cf45-44ec-a2a5-f7210ba72289" containerName="nova-scheduler-scheduler" containerID="cri-o://9b8408fdb07e2e8da5bb7a92bc9bfa21a078482b939c9cf3d3ce6d013498d2c7" gracePeriod=30 Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.686525 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": EOF" Oct 04 04:59:23 crc kubenswrapper[4575]: I1004 04:59:23.686566 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": EOF" Oct 04 04:59:24 crc kubenswrapper[4575]: I1004 04:59:24.388379 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4ml2t" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:24 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:24 crc kubenswrapper[4575]: > Oct 04 04:59:24 crc kubenswrapper[4575]: I1004 04:59:24.510273 4575 generic.go:334] "Generic (PLEG): container finished" podID="8b477b40-b8de-400b-9caa-e38540b59c47" containerID="f61d6d584e189139f2e51e9ce576f8d5d56641866bfbf7a2dd52539b5a2a704a" exitCode=143 Oct 04 04:59:24 crc kubenswrapper[4575]: I1004 04:59:24.510329 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8b477b40-b8de-400b-9caa-e38540b59c47","Type":"ContainerDied","Data":"f61d6d584e189139f2e51e9ce576f8d5d56641866bfbf7a2dd52539b5a2a704a"} Oct 04 04:59:24 crc kubenswrapper[4575]: I1004 04:59:24.983761 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.150314 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-scripts\") pod \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.150397 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbcqj\" (UniqueName: \"kubernetes.io/projected/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-kube-api-access-tbcqj\") pod \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.150465 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-config-data\") pod \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.150563 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-combined-ca-bundle\") pod \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\" (UID: \"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74\") " Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.156951 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-scripts" (OuterVolumeSpecName: "scripts") pod "6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74" (UID: "6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.161874 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-kube-api-access-tbcqj" (OuterVolumeSpecName: "kube-api-access-tbcqj") pod "6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74" (UID: "6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74"). InnerVolumeSpecName "kube-api-access-tbcqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.185053 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74" (UID: "6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.188862 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-config-data" (OuterVolumeSpecName: "config-data") pod "6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74" (UID: "6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.252879 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.252924 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbcqj\" (UniqueName: \"kubernetes.io/projected/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-kube-api-access-tbcqj\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.252940 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.252951 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.521525 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-znqtq" event={"ID":"6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74","Type":"ContainerDied","Data":"f5734b42efd5abbc026d4ece064b6dbde6f28cddc6263f0765661609eb85dd16"} Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.521870 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5734b42efd5abbc026d4ece064b6dbde6f28cddc6263f0765661609eb85dd16" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.521637 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-znqtq" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.612830 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:59:25 crc kubenswrapper[4575]: E1004 04:59:25.613370 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8" containerName="nova-manage" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.613395 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8" containerName="nova-manage" Oct 04 04:59:25 crc kubenswrapper[4575]: E1004 04:59:25.613432 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74" containerName="nova-cell1-conductor-db-sync" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.613440 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74" containerName="nova-cell1-conductor-db-sync" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.613679 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74" containerName="nova-cell1-conductor-db-sync" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.613724 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8" containerName="nova-manage" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.614357 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.617761 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.633175 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.767894 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5823bcd4-8223-4b61-a2a3-aa5cea21fc3b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.767948 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz6c7\" (UniqueName: \"kubernetes.io/projected/5823bcd4-8223-4b61-a2a3-aa5cea21fc3b-kube-api-access-xz6c7\") pod \"nova-cell1-conductor-0\" (UID: \"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.767971 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5823bcd4-8223-4b61-a2a3-aa5cea21fc3b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.899056 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5823bcd4-8223-4b61-a2a3-aa5cea21fc3b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.899132 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz6c7\" (UniqueName: \"kubernetes.io/projected/5823bcd4-8223-4b61-a2a3-aa5cea21fc3b-kube-api-access-xz6c7\") pod \"nova-cell1-conductor-0\" (UID: \"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.899167 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5823bcd4-8223-4b61-a2a3-aa5cea21fc3b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.911103 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5823bcd4-8223-4b61-a2a3-aa5cea21fc3b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.915536 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5823bcd4-8223-4b61-a2a3-aa5cea21fc3b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.927711 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz6c7\" (UniqueName: \"kubernetes.io/projected/5823bcd4-8223-4b61-a2a3-aa5cea21fc3b-kube-api-access-xz6c7\") pod \"nova-cell1-conductor-0\" (UID: \"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:25 crc kubenswrapper[4575]: I1004 04:59:25.932709 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:26 crc kubenswrapper[4575]: I1004 04:59:26.506218 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:59:26 crc kubenswrapper[4575]: I1004 04:59:26.537426 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b","Type":"ContainerStarted","Data":"e986bb5ccc855b9108f02d5be46a5f2a2863cce7276717655c174e3388a0dcd8"} Oct 04 04:59:26 crc kubenswrapper[4575]: I1004 04:59:26.545335 4575 generic.go:334] "Generic (PLEG): container finished" podID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerID="376a8e8470b9959c733967cacb7a6d6b2b157401d6330173cf7a8725128fbf8f" exitCode=0 Oct 04 04:59:26 crc kubenswrapper[4575]: I1004 04:59:26.545439 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerDied","Data":"376a8e8470b9959c733967cacb7a6d6b2b157401d6330173cf7a8725128fbf8f"} Oct 04 04:59:26 crc kubenswrapper[4575]: I1004 04:59:26.545476 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerStarted","Data":"1286c3d99af99d1bd1f66544432932ee55a667d406d520ce4179dfad08a0c438"} Oct 04 04:59:26 crc kubenswrapper[4575]: I1004 04:59:26.545527 4575 scope.go:117] "RemoveContainer" containerID="8d0029179c5592b7aefb48edf15d738cddb00579282fc0b3c8d4eb772010beee" Oct 04 04:59:27 crc kubenswrapper[4575]: I1004 04:59:27.288968 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:27 crc kubenswrapper[4575]: I1004 04:59:27.289054 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:27 crc kubenswrapper[4575]: I1004 04:59:27.555834 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5823bcd4-8223-4b61-a2a3-aa5cea21fc3b","Type":"ContainerStarted","Data":"865152a929d55acef40dbc988c9fb707ff7024a949ef5a10b434812601136b78"} Oct 04 04:59:27 crc kubenswrapper[4575]: I1004 04:59:27.556857 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:27 crc kubenswrapper[4575]: I1004 04:59:27.599803 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.599778137 podStartE2EDuration="2.599778137s" podCreationTimestamp="2025-10-04 04:59:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:27.590238862 +0000 UTC m=+1518.918797686" watchObservedRunningTime="2025-10-04 04:59:27.599778137 +0000 UTC m=+1518.928336951" Oct 04 04:59:27 crc kubenswrapper[4575]: E1004 04:59:27.613185 4575 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b8408fdb07e2e8da5bb7a92bc9bfa21a078482b939c9cf3d3ce6d013498d2c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:59:27 crc kubenswrapper[4575]: E1004 04:59:27.614867 4575 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b8408fdb07e2e8da5bb7a92bc9bfa21a078482b939c9cf3d3ce6d013498d2c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:59:27 crc kubenswrapper[4575]: E1004 04:59:27.619410 4575 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b8408fdb07e2e8da5bb7a92bc9bfa21a078482b939c9cf3d3ce6d013498d2c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:59:27 crc kubenswrapper[4575]: E1004 04:59:27.619466 4575 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="75082b13-cf45-44ec-a2a5-f7210ba72289" containerName="nova-scheduler-scheduler" Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.377252 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-qj7qr" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:28 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:28 crc kubenswrapper[4575]: > Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.568747 4575 generic.go:334] "Generic (PLEG): container finished" podID="75082b13-cf45-44ec-a2a5-f7210ba72289" containerID="9b8408fdb07e2e8da5bb7a92bc9bfa21a078482b939c9cf3d3ce6d013498d2c7" exitCode=0 Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.568819 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"75082b13-cf45-44ec-a2a5-f7210ba72289","Type":"ContainerDied","Data":"9b8408fdb07e2e8da5bb7a92bc9bfa21a078482b939c9cf3d3ce6d013498d2c7"} Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.568853 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"75082b13-cf45-44ec-a2a5-f7210ba72289","Type":"ContainerDied","Data":"6ff5c863d1d316972c2d519ad205f6dab0057562fc129e27dcc442e8677a3141"} Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.568864 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ff5c863d1d316972c2d519ad205f6dab0057562fc129e27dcc442e8677a3141" Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.630296 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.759048 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-config-data\") pod \"75082b13-cf45-44ec-a2a5-f7210ba72289\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.759116 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-combined-ca-bundle\") pod \"75082b13-cf45-44ec-a2a5-f7210ba72289\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.759202 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2fgr\" (UniqueName: \"kubernetes.io/projected/75082b13-cf45-44ec-a2a5-f7210ba72289-kube-api-access-x2fgr\") pod \"75082b13-cf45-44ec-a2a5-f7210ba72289\" (UID: \"75082b13-cf45-44ec-a2a5-f7210ba72289\") " Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.768883 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75082b13-cf45-44ec-a2a5-f7210ba72289-kube-api-access-x2fgr" (OuterVolumeSpecName: "kube-api-access-x2fgr") pod "75082b13-cf45-44ec-a2a5-f7210ba72289" (UID: "75082b13-cf45-44ec-a2a5-f7210ba72289"). InnerVolumeSpecName "kube-api-access-x2fgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.792654 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75082b13-cf45-44ec-a2a5-f7210ba72289" (UID: "75082b13-cf45-44ec-a2a5-f7210ba72289"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.797784 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-config-data" (OuterVolumeSpecName: "config-data") pod "75082b13-cf45-44ec-a2a5-f7210ba72289" (UID: "75082b13-cf45-44ec-a2a5-f7210ba72289"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.861816 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.861858 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75082b13-cf45-44ec-a2a5-f7210ba72289-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:28 crc kubenswrapper[4575]: I1004 04:59:28.861874 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2fgr\" (UniqueName: \"kubernetes.io/projected/75082b13-cf45-44ec-a2a5-f7210ba72289-kube-api-access-x2fgr\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.013165 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.577462 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.656432 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.683283 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.698737 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:59:29 crc kubenswrapper[4575]: E1004 04:59:29.699233 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75082b13-cf45-44ec-a2a5-f7210ba72289" containerName="nova-scheduler-scheduler" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.699253 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="75082b13-cf45-44ec-a2a5-f7210ba72289" containerName="nova-scheduler-scheduler" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.699532 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="75082b13-cf45-44ec-a2a5-f7210ba72289" containerName="nova-scheduler-scheduler" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.700370 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.707167 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.711670 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.882921 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.883325 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-config-data\") pod \"nova-scheduler-0\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.883529 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65xnk\" (UniqueName: \"kubernetes.io/projected/15a3821f-f7ca-45b4-b667-d5985bc25f3a-kube-api-access-65xnk\") pod \"nova-scheduler-0\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.985077 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.985136 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-config-data\") pod \"nova-scheduler-0\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.985294 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65xnk\" (UniqueName: \"kubernetes.io/projected/15a3821f-f7ca-45b4-b667-d5985bc25f3a-kube-api-access-65xnk\") pod \"nova-scheduler-0\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.993474 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-config-data\") pod \"nova-scheduler-0\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:29 crc kubenswrapper[4575]: I1004 04:59:29.997657 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:30 crc kubenswrapper[4575]: I1004 04:59:30.003610 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65xnk\" (UniqueName: \"kubernetes.io/projected/15a3821f-f7ca-45b4-b667-d5985bc25f3a-kube-api-access-65xnk\") pod \"nova-scheduler-0\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " pod="openstack/nova-scheduler-0" Oct 04 04:59:30 crc kubenswrapper[4575]: I1004 04:59:30.035318 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:59:30 crc kubenswrapper[4575]: I1004 04:59:30.521087 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtc62" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:30 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:30 crc kubenswrapper[4575]: > Oct 04 04:59:30 crc kubenswrapper[4575]: I1004 04:59:30.622677 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:59:30 crc kubenswrapper[4575]: W1004 04:59:30.646016 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15a3821f_f7ca_45b4_b667_d5985bc25f3a.slice/crio-bd87db666666b645584df8cbb413b693281a018585f08eaa9bbe2522ee052e07 WatchSource:0}: Error finding container bd87db666666b645584df8cbb413b693281a018585f08eaa9bbe2522ee052e07: Status 404 returned error can't find the container with id bd87db666666b645584df8cbb413b693281a018585f08eaa9bbe2522ee052e07 Oct 04 04:59:30 crc kubenswrapper[4575]: I1004 04:59:30.719746 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:59:30 crc kubenswrapper[4575]: I1004 04:59:30.719850 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:59:30 crc kubenswrapper[4575]: I1004 04:59:30.720764 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"ec938e12e0dcd7b18a29f344980b6af8ff56070e067b69f96ecbd149ba683cb0"} pod="openstack/horizon-644bf5cdd4-pwdw9" containerMessage="Container horizon failed startup probe, will be restarted" Oct 04 04:59:30 crc kubenswrapper[4575]: I1004 04:59:30.720807 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" containerID="cri-o://ec938e12e0dcd7b18a29f344980b6af8ff56070e067b69f96ecbd149ba683cb0" gracePeriod=30 Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.323504 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75082b13-cf45-44ec-a2a5-f7210ba72289" path="/var/lib/kubelet/pods/75082b13-cf45-44ec-a2a5-f7210ba72289/volumes" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.598575 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15a3821f-f7ca-45b4-b667-d5985bc25f3a","Type":"ContainerStarted","Data":"c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c"} Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.598872 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15a3821f-f7ca-45b4-b667-d5985bc25f3a","Type":"ContainerStarted","Data":"bd87db666666b645584df8cbb413b693281a018585f08eaa9bbe2522ee052e07"} Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.600891 4575 generic.go:334] "Generic (PLEG): container finished" podID="8b477b40-b8de-400b-9caa-e38540b59c47" containerID="865c212f049bf39aec2d2c254afacf51d842fae1342f99c2bf42dbfc9f22dbb8" exitCode=0 Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.600928 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8b477b40-b8de-400b-9caa-e38540b59c47","Type":"ContainerDied","Data":"865c212f049bf39aec2d2c254afacf51d842fae1342f99c2bf42dbfc9f22dbb8"} Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.600949 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8b477b40-b8de-400b-9caa-e38540b59c47","Type":"ContainerDied","Data":"50e8c4ea93e8cb6db66947a2356f61f4d37da1467e80737faa491029afa0e025"} Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.600960 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50e8c4ea93e8cb6db66947a2356f61f4d37da1467e80737faa491029afa0e025" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.621335 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.621315014 podStartE2EDuration="2.621315014s" podCreationTimestamp="2025-10-04 04:59:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:31.615768574 +0000 UTC m=+1522.944327388" watchObservedRunningTime="2025-10-04 04:59:31.621315014 +0000 UTC m=+1522.949873828" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.675023 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.724618 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-config-data\") pod \"8b477b40-b8de-400b-9caa-e38540b59c47\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.724671 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd55d\" (UniqueName: \"kubernetes.io/projected/8b477b40-b8de-400b-9caa-e38540b59c47-kube-api-access-jd55d\") pod \"8b477b40-b8de-400b-9caa-e38540b59c47\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.725298 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b477b40-b8de-400b-9caa-e38540b59c47-logs\") pod \"8b477b40-b8de-400b-9caa-e38540b59c47\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.725334 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b477b40-b8de-400b-9caa-e38540b59c47-logs" (OuterVolumeSpecName: "logs") pod "8b477b40-b8de-400b-9caa-e38540b59c47" (UID: "8b477b40-b8de-400b-9caa-e38540b59c47"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.725356 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-combined-ca-bundle\") pod \"8b477b40-b8de-400b-9caa-e38540b59c47\" (UID: \"8b477b40-b8de-400b-9caa-e38540b59c47\") " Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.725957 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b477b40-b8de-400b-9caa-e38540b59c47-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.749841 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b477b40-b8de-400b-9caa-e38540b59c47-kube-api-access-jd55d" (OuterVolumeSpecName: "kube-api-access-jd55d") pod "8b477b40-b8de-400b-9caa-e38540b59c47" (UID: "8b477b40-b8de-400b-9caa-e38540b59c47"). InnerVolumeSpecName "kube-api-access-jd55d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.770846 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b477b40-b8de-400b-9caa-e38540b59c47" (UID: "8b477b40-b8de-400b-9caa-e38540b59c47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.775458 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-config-data" (OuterVolumeSpecName: "config-data") pod "8b477b40-b8de-400b-9caa-e38540b59c47" (UID: "8b477b40-b8de-400b-9caa-e38540b59c47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.828605 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.828639 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd55d\" (UniqueName: \"kubernetes.io/projected/8b477b40-b8de-400b-9caa-e38540b59c47-kube-api-access-jd55d\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:31 crc kubenswrapper[4575]: I1004 04:59:31.828656 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b477b40-b8de-400b-9caa-e38540b59c47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.611184 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.655815 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.667090 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.689259 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:32 crc kubenswrapper[4575]: E1004 04:59:32.691026 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-log" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.691142 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-log" Oct 04 04:59:32 crc kubenswrapper[4575]: E1004 04:59:32.691230 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-api" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.691306 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-api" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.691715 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-api" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.691805 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" containerName="nova-api-log" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.693263 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.701290 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.701944 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.775833 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-config-data\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.775934 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8a470d-7d56-4511-b59a-3c4063676ba8-logs\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.776003 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.776046 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kw2v\" (UniqueName: \"kubernetes.io/projected/cf8a470d-7d56-4511-b59a-3c4063676ba8-kube-api-access-5kw2v\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.877735 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kw2v\" (UniqueName: \"kubernetes.io/projected/cf8a470d-7d56-4511-b59a-3c4063676ba8-kube-api-access-5kw2v\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.877888 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-config-data\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.877921 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8a470d-7d56-4511-b59a-3c4063676ba8-logs\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.877999 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.878714 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8a470d-7d56-4511-b59a-3c4063676ba8-logs\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.884564 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.897009 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-config-data\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:32 crc kubenswrapper[4575]: I1004 04:59:32.899935 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kw2v\" (UniqueName: \"kubernetes.io/projected/cf8a470d-7d56-4511-b59a-3c4063676ba8-kube-api-access-5kw2v\") pod \"nova-api-0\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " pod="openstack/nova-api-0" Oct 04 04:59:33 crc kubenswrapper[4575]: I1004 04:59:33.013565 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:59:33 crc kubenswrapper[4575]: I1004 04:59:33.337125 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b477b40-b8de-400b-9caa-e38540b59c47" path="/var/lib/kubelet/pods/8b477b40-b8de-400b-9caa-e38540b59c47/volumes" Oct 04 04:59:33 crc kubenswrapper[4575]: I1004 04:59:33.569500 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:33 crc kubenswrapper[4575]: I1004 04:59:33.633488 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf8a470d-7d56-4511-b59a-3c4063676ba8","Type":"ContainerStarted","Data":"32596259c505ab716d5cd86f58654c90bd0f5bf9e3b83ccd4758d10809202430"} Oct 04 04:59:34 crc kubenswrapper[4575]: I1004 04:59:34.377016 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4ml2t" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:34 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:34 crc kubenswrapper[4575]: > Oct 04 04:59:34 crc kubenswrapper[4575]: I1004 04:59:34.663763 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf8a470d-7d56-4511-b59a-3c4063676ba8","Type":"ContainerStarted","Data":"2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4"} Oct 04 04:59:34 crc kubenswrapper[4575]: I1004 04:59:34.663817 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf8a470d-7d56-4511-b59a-3c4063676ba8","Type":"ContainerStarted","Data":"72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a"} Oct 04 04:59:34 crc kubenswrapper[4575]: I1004 04:59:34.695188 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.695159812 podStartE2EDuration="2.695159812s" podCreationTimestamp="2025-10-04 04:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:34.684459643 +0000 UTC m=+1526.013018467" watchObservedRunningTime="2025-10-04 04:59:34.695159812 +0000 UTC m=+1526.023718626" Oct 04 04:59:34 crc kubenswrapper[4575]: I1004 04:59:34.884691 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:59:34 crc kubenswrapper[4575]: I1004 04:59:34.885127 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="010127df-25eb-4251-91a4-5cf00fafc403" containerName="kube-state-metrics" containerID="cri-o://79b8cb3b19aab4a1345eb2ddc54f6a6af93bcc6bdd6b18c48a8aace11cf3b3af" gracePeriod=30 Oct 04 04:59:35 crc kubenswrapper[4575]: I1004 04:59:35.036185 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:59:35 crc kubenswrapper[4575]: I1004 04:59:35.559729 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:59:35 crc kubenswrapper[4575]: I1004 04:59:35.559783 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 04:59:35 crc kubenswrapper[4575]: I1004 04:59:35.681538 4575 generic.go:334] "Generic (PLEG): container finished" podID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerID="ec938e12e0dcd7b18a29f344980b6af8ff56070e067b69f96ecbd149ba683cb0" exitCode=0 Oct 04 04:59:35 crc kubenswrapper[4575]: I1004 04:59:35.682792 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerDied","Data":"ec938e12e0dcd7b18a29f344980b6af8ff56070e067b69f96ecbd149ba683cb0"} Oct 04 04:59:35 crc kubenswrapper[4575]: I1004 04:59:35.682856 4575 scope.go:117] "RemoveContainer" containerID="e66aa0b86f8f6573106bccc2c81186d127b7315b211382f7d35959f05ac6d4fb" Oct 04 04:59:35 crc kubenswrapper[4575]: I1004 04:59:35.685869 4575 generic.go:334] "Generic (PLEG): container finished" podID="010127df-25eb-4251-91a4-5cf00fafc403" containerID="79b8cb3b19aab4a1345eb2ddc54f6a6af93bcc6bdd6b18c48a8aace11cf3b3af" exitCode=2 Oct 04 04:59:35 crc kubenswrapper[4575]: I1004 04:59:35.687282 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"010127df-25eb-4251-91a4-5cf00fafc403","Type":"ContainerDied","Data":"79b8cb3b19aab4a1345eb2ddc54f6a6af93bcc6bdd6b18c48a8aace11cf3b3af"} Oct 04 04:59:35 crc kubenswrapper[4575]: I1004 04:59:35.979248 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.060666 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.262778 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfpvw\" (UniqueName: \"kubernetes.io/projected/010127df-25eb-4251-91a4-5cf00fafc403-kube-api-access-zfpvw\") pod \"010127df-25eb-4251-91a4-5cf00fafc403\" (UID: \"010127df-25eb-4251-91a4-5cf00fafc403\") " Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.273833 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/010127df-25eb-4251-91a4-5cf00fafc403-kube-api-access-zfpvw" (OuterVolumeSpecName: "kube-api-access-zfpvw") pod "010127df-25eb-4251-91a4-5cf00fafc403" (UID: "010127df-25eb-4251-91a4-5cf00fafc403"). InnerVolumeSpecName "kube-api-access-zfpvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.366758 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfpvw\" (UniqueName: \"kubernetes.io/projected/010127df-25eb-4251-91a4-5cf00fafc403-kube-api-access-zfpvw\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.701480 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerStarted","Data":"f8d4443c36b0c2245629253c64c8178d2ca6381e985a3cf457f5457982a4dcd8"} Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.704600 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"010127df-25eb-4251-91a4-5cf00fafc403","Type":"ContainerDied","Data":"0d65d6abea2c703a5bd842b96eaf3ca4fb631ec2fc1a22139f16ff853a50ac8e"} Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.704750 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.704755 4575 scope.go:117] "RemoveContainer" containerID="79b8cb3b19aab4a1345eb2ddc54f6a6af93bcc6bdd6b18c48a8aace11cf3b3af" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.794747 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.821684 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.834041 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:59:36 crc kubenswrapper[4575]: E1004 04:59:36.834573 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="010127df-25eb-4251-91a4-5cf00fafc403" containerName="kube-state-metrics" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.834609 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="010127df-25eb-4251-91a4-5cf00fafc403" containerName="kube-state-metrics" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.834817 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="010127df-25eb-4251-91a4-5cf00fafc403" containerName="kube-state-metrics" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.835535 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.841690 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.842350 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.842332 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.981524 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1a6e2274-1cf7-4762-a0a3-bec545e642c8-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.981631 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a6e2274-1cf7-4762-a0a3-bec545e642c8-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.981682 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsqg2\" (UniqueName: \"kubernetes.io/projected/1a6e2274-1cf7-4762-a0a3-bec545e642c8-kube-api-access-jsqg2\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.981786 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a6e2274-1cf7-4762-a0a3-bec545e642c8-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.993029 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.993305 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="ceilometer-central-agent" containerID="cri-o://7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0" gracePeriod=30 Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.993420 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="ceilometer-notification-agent" containerID="cri-o://a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9" gracePeriod=30 Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.993423 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="sg-core" containerID="cri-o://e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1" gracePeriod=30 Oct 04 04:59:36 crc kubenswrapper[4575]: I1004 04:59:36.993469 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="proxy-httpd" containerID="cri-o://1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876" gracePeriod=30 Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.083603 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1a6e2274-1cf7-4762-a0a3-bec545e642c8-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.083712 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a6e2274-1cf7-4762-a0a3-bec545e642c8-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.083765 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsqg2\" (UniqueName: \"kubernetes.io/projected/1a6e2274-1cf7-4762-a0a3-bec545e642c8-kube-api-access-jsqg2\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.083859 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a6e2274-1cf7-4762-a0a3-bec545e642c8-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.089272 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1a6e2274-1cf7-4762-a0a3-bec545e642c8-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.090204 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a6e2274-1cf7-4762-a0a3-bec545e642c8-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.097190 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a6e2274-1cf7-4762-a0a3-bec545e642c8-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.113651 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsqg2\" (UniqueName: \"kubernetes.io/projected/1a6e2274-1cf7-4762-a0a3-bec545e642c8-kube-api-access-jsqg2\") pod \"kube-state-metrics-0\" (UID: \"1a6e2274-1cf7-4762-a0a3-bec545e642c8\") " pod="openstack/kube-state-metrics-0" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.160164 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.365399 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="010127df-25eb-4251-91a4-5cf00fafc403" path="/var/lib/kubelet/pods/010127df-25eb-4251-91a4-5cf00fafc403/volumes" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.398003 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.538848 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.652671 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qj7qr"] Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.716845 4575 generic.go:334] "Generic (PLEG): container finished" podID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerID="1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876" exitCode=0 Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.716875 4575 generic.go:334] "Generic (PLEG): container finished" podID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerID="e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1" exitCode=2 Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.717622 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerDied","Data":"1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876"} Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.717644 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerDied","Data":"e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1"} Oct 04 04:59:37 crc kubenswrapper[4575]: I1004 04:59:37.774296 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:59:37 crc kubenswrapper[4575]: W1004 04:59:37.781151 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a6e2274_1cf7_4762_a0a3_bec545e642c8.slice/crio-e7ffe5b966315674cc97caa26c79083c2f38fd84f308b3f83f8de825967ead68 WatchSource:0}: Error finding container e7ffe5b966315674cc97caa26c79083c2f38fd84f308b3f83f8de825967ead68: Status 404 returned error can't find the container with id e7ffe5b966315674cc97caa26c79083c2f38fd84f308b3f83f8de825967ead68 Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.715356 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.727927 4575 generic.go:334] "Generic (PLEG): container finished" podID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerID="a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9" exitCode=0 Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.727959 4575 generic.go:334] "Generic (PLEG): container finished" podID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerID="7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0" exitCode=0 Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.728004 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerDied","Data":"a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9"} Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.728034 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerDied","Data":"7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0"} Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.728045 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26a904de-6d26-41ce-a29b-7cd0dbe0f610","Type":"ContainerDied","Data":"c85851676b289c4e7eb2c406a43570e6af7dfe4c5e3a6de3ed9aa1b071e149f8"} Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.728059 4575 scope.go:117] "RemoveContainer" containerID="1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.728169 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.733912 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-combined-ca-bundle\") pod \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.733968 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-scripts\") pod \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.734001 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-run-httpd\") pod \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.734048 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-sg-core-conf-yaml\") pod \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.734930 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "26a904de-6d26-41ce-a29b-7cd0dbe0f610" (UID: "26a904de-6d26-41ce-a29b-7cd0dbe0f610"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.735203 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1a6e2274-1cf7-4762-a0a3-bec545e642c8","Type":"ContainerStarted","Data":"14a56ef8c2043b655a5284a959f5f86a774bb74123675143ad9745c4076f4054"} Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.735264 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qj7qr" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerName="registry-server" containerID="cri-o://a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c" gracePeriod=2 Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.735276 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1a6e2274-1cf7-4762-a0a3-bec545e642c8","Type":"ContainerStarted","Data":"e7ffe5b966315674cc97caa26c79083c2f38fd84f308b3f83f8de825967ead68"} Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.735748 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.743612 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-scripts" (OuterVolumeSpecName: "scripts") pod "26a904de-6d26-41ce-a29b-7cd0dbe0f610" (UID: "26a904de-6d26-41ce-a29b-7cd0dbe0f610"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.815368 4575 scope.go:117] "RemoveContainer" containerID="e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.822347 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.369999337 podStartE2EDuration="2.822246123s" podCreationTimestamp="2025-10-04 04:59:36 +0000 UTC" firstStartedPulling="2025-10-04 04:59:37.783265572 +0000 UTC m=+1529.111824386" lastFinishedPulling="2025-10-04 04:59:38.235512358 +0000 UTC m=+1529.564071172" observedRunningTime="2025-10-04 04:59:38.811789881 +0000 UTC m=+1530.140348695" watchObservedRunningTime="2025-10-04 04:59:38.822246123 +0000 UTC m=+1530.150804947" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.836256 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2kgn\" (UniqueName: \"kubernetes.io/projected/26a904de-6d26-41ce-a29b-7cd0dbe0f610-kube-api-access-c2kgn\") pod \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.836339 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-log-httpd\") pod \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.836368 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-config-data\") pod \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\" (UID: \"26a904de-6d26-41ce-a29b-7cd0dbe0f610\") " Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.837021 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.837048 4575 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.837309 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "26a904de-6d26-41ce-a29b-7cd0dbe0f610" (UID: "26a904de-6d26-41ce-a29b-7cd0dbe0f610"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.838451 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "26a904de-6d26-41ce-a29b-7cd0dbe0f610" (UID: "26a904de-6d26-41ce-a29b-7cd0dbe0f610"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.850157 4575 scope.go:117] "RemoveContainer" containerID="a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.854325 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a904de-6d26-41ce-a29b-7cd0dbe0f610-kube-api-access-c2kgn" (OuterVolumeSpecName: "kube-api-access-c2kgn") pod "26a904de-6d26-41ce-a29b-7cd0dbe0f610" (UID: "26a904de-6d26-41ce-a29b-7cd0dbe0f610"). InnerVolumeSpecName "kube-api-access-c2kgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.876799 4575 scope.go:117] "RemoveContainer" containerID="7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.908539 4575 scope.go:117] "RemoveContainer" containerID="1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876" Oct 04 04:59:38 crc kubenswrapper[4575]: E1004 04:59:38.909325 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876\": container with ID starting with 1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876 not found: ID does not exist" containerID="1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.909374 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876"} err="failed to get container status \"1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876\": rpc error: code = NotFound desc = could not find container \"1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876\": container with ID starting with 1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876 not found: ID does not exist" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.909401 4575 scope.go:117] "RemoveContainer" containerID="e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1" Oct 04 04:59:38 crc kubenswrapper[4575]: E1004 04:59:38.910018 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1\": container with ID starting with e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1 not found: ID does not exist" containerID="e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.910054 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1"} err="failed to get container status \"e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1\": rpc error: code = NotFound desc = could not find container \"e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1\": container with ID starting with e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1 not found: ID does not exist" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.910071 4575 scope.go:117] "RemoveContainer" containerID="a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9" Oct 04 04:59:38 crc kubenswrapper[4575]: E1004 04:59:38.910297 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9\": container with ID starting with a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9 not found: ID does not exist" containerID="a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.910326 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9"} err="failed to get container status \"a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9\": rpc error: code = NotFound desc = could not find container \"a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9\": container with ID starting with a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9 not found: ID does not exist" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.910344 4575 scope.go:117] "RemoveContainer" containerID="7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0" Oct 04 04:59:38 crc kubenswrapper[4575]: E1004 04:59:38.910548 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0\": container with ID starting with 7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0 not found: ID does not exist" containerID="7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.910577 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0"} err="failed to get container status \"7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0\": rpc error: code = NotFound desc = could not find container \"7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0\": container with ID starting with 7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0 not found: ID does not exist" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.910611 4575 scope.go:117] "RemoveContainer" containerID="1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.910811 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876"} err="failed to get container status \"1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876\": rpc error: code = NotFound desc = could not find container \"1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876\": container with ID starting with 1c1bf05c5443fe8bc5efbe1742aa170c41e85245b06dbdf88b6b93b951dc1876 not found: ID does not exist" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.910837 4575 scope.go:117] "RemoveContainer" containerID="e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.911042 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1"} err="failed to get container status \"e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1\": rpc error: code = NotFound desc = could not find container \"e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1\": container with ID starting with e1c61f21781a46e10856684610e98785d487c8e084c1f29efb7baee18001afa1 not found: ID does not exist" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.911068 4575 scope.go:117] "RemoveContainer" containerID="a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.911257 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9"} err="failed to get container status \"a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9\": rpc error: code = NotFound desc = could not find container \"a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9\": container with ID starting with a37d6271c76935e957ff06c6548ac9be785456c151c5beb847f09a7026e6d7e9 not found: ID does not exist" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.911286 4575 scope.go:117] "RemoveContainer" containerID="7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.911475 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0"} err="failed to get container status \"7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0\": rpc error: code = NotFound desc = could not find container \"7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0\": container with ID starting with 7d42c70ee3aa51d6e0f22b65a3014656738a1b3db9a4291f9457b19d1ebbf2f0 not found: ID does not exist" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.916750 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26a904de-6d26-41ce-a29b-7cd0dbe0f610" (UID: "26a904de-6d26-41ce-a29b-7cd0dbe0f610"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.939008 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.939041 4575 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.939066 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2kgn\" (UniqueName: \"kubernetes.io/projected/26a904de-6d26-41ce-a29b-7cd0dbe0f610-kube-api-access-c2kgn\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.939079 4575 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26a904de-6d26-41ce-a29b-7cd0dbe0f610-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:38 crc kubenswrapper[4575]: I1004 04:59:38.966652 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-config-data" (OuterVolumeSpecName: "config-data") pod "26a904de-6d26-41ce-a29b-7cd0dbe0f610" (UID: "26a904de-6d26-41ce-a29b-7cd0dbe0f610"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.040707 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a904de-6d26-41ce-a29b-7cd0dbe0f610-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.079320 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.102049 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.114376 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:39 crc kubenswrapper[4575]: E1004 04:59:39.114900 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="proxy-httpd" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.114924 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="proxy-httpd" Oct 04 04:59:39 crc kubenswrapper[4575]: E1004 04:59:39.114956 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="sg-core" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.114964 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="sg-core" Oct 04 04:59:39 crc kubenswrapper[4575]: E1004 04:59:39.114975 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="ceilometer-central-agent" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.114984 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="ceilometer-central-agent" Oct 04 04:59:39 crc kubenswrapper[4575]: E1004 04:59:39.114995 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="ceilometer-notification-agent" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.115003 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="ceilometer-notification-agent" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.115240 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="sg-core" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.115255 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="proxy-httpd" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.115281 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="ceilometer-central-agent" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.115297 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" containerName="ceilometer-notification-agent" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.117478 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.122355 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.122615 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.122754 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.134976 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.146852 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.146888 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-log-httpd\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.146916 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-scripts\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.146945 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.146974 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.146999 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-run-httpd\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.147026 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-config-data\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.147057 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6czkd\" (UniqueName: \"kubernetes.io/projected/5bc4c7a4-dd7f-4e32-aebe-79342e605977-kube-api-access-6czkd\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.248397 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-config-data\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.248476 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6czkd\" (UniqueName: \"kubernetes.io/projected/5bc4c7a4-dd7f-4e32-aebe-79342e605977-kube-api-access-6czkd\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.248578 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.248724 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-log-httpd\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.248760 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-scripts\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.248799 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.248856 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.248897 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-run-httpd\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.249420 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-run-httpd\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.249706 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-log-httpd\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.255564 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.255669 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.256296 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-scripts\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.257116 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-config-data\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.267562 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.271398 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6czkd\" (UniqueName: \"kubernetes.io/projected/5bc4c7a4-dd7f-4e32-aebe-79342e605977-kube-api-access-6czkd\") pod \"ceilometer-0\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.355511 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26a904de-6d26-41ce-a29b-7cd0dbe0f610" path="/var/lib/kubelet/pods/26a904de-6d26-41ce-a29b-7cd0dbe0f610/volumes" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.491969 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.723036 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.810852 4575 generic.go:334] "Generic (PLEG): container finished" podID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerID="a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c" exitCode=0 Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.810930 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qj7qr" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.810952 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj7qr" event={"ID":"2b91f552-6eb5-479d-9a79-07a356d33ab0","Type":"ContainerDied","Data":"a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c"} Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.812971 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj7qr" event={"ID":"2b91f552-6eb5-479d-9a79-07a356d33ab0","Type":"ContainerDied","Data":"67007a30e75dbb01dfaf2954e83a968bd59ba71005fd6aac4acdc7055fab3898"} Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.812995 4575 scope.go:117] "RemoveContainer" containerID="a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.869571 4575 scope.go:117] "RemoveContainer" containerID="3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.871264 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74bsl\" (UniqueName: \"kubernetes.io/projected/2b91f552-6eb5-479d-9a79-07a356d33ab0-kube-api-access-74bsl\") pod \"2b91f552-6eb5-479d-9a79-07a356d33ab0\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.872474 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-catalog-content\") pod \"2b91f552-6eb5-479d-9a79-07a356d33ab0\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.872662 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-utilities\") pod \"2b91f552-6eb5-479d-9a79-07a356d33ab0\" (UID: \"2b91f552-6eb5-479d-9a79-07a356d33ab0\") " Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.874363 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-utilities" (OuterVolumeSpecName: "utilities") pod "2b91f552-6eb5-479d-9a79-07a356d33ab0" (UID: "2b91f552-6eb5-479d-9a79-07a356d33ab0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.875067 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.899943 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b91f552-6eb5-479d-9a79-07a356d33ab0-kube-api-access-74bsl" (OuterVolumeSpecName: "kube-api-access-74bsl") pod "2b91f552-6eb5-479d-9a79-07a356d33ab0" (UID: "2b91f552-6eb5-479d-9a79-07a356d33ab0"). InnerVolumeSpecName "kube-api-access-74bsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.937031 4575 scope.go:117] "RemoveContainer" containerID="a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.976550 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74bsl\" (UniqueName: \"kubernetes.io/projected/2b91f552-6eb5-479d-9a79-07a356d33ab0-kube-api-access-74bsl\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.977014 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b91f552-6eb5-479d-9a79-07a356d33ab0" (UID: "2b91f552-6eb5-479d-9a79-07a356d33ab0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.985274 4575 scope.go:117] "RemoveContainer" containerID="a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c" Oct 04 04:59:39 crc kubenswrapper[4575]: E1004 04:59:39.987122 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c\": container with ID starting with a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c not found: ID does not exist" containerID="a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.987163 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c"} err="failed to get container status \"a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c\": rpc error: code = NotFound desc = could not find container \"a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c\": container with ID starting with a9378acd9eff78c588c1c8fc04f88744181e2535cd794729ceb20ee91a34fa3c not found: ID does not exist" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.987191 4575 scope.go:117] "RemoveContainer" containerID="3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989" Oct 04 04:59:39 crc kubenswrapper[4575]: E1004 04:59:39.989151 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989\": container with ID starting with 3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989 not found: ID does not exist" containerID="3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.989190 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989"} err="failed to get container status \"3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989\": rpc error: code = NotFound desc = could not find container \"3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989\": container with ID starting with 3bef8cf3272cc7736d67944f31d838c22da98359103b121c4f03b3814eb73989 not found: ID does not exist" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.989213 4575 scope.go:117] "RemoveContainer" containerID="a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a" Oct 04 04:59:39 crc kubenswrapper[4575]: E1004 04:59:39.989554 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a\": container with ID starting with a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a not found: ID does not exist" containerID="a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a" Oct 04 04:59:39 crc kubenswrapper[4575]: I1004 04:59:39.989579 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a"} err="failed to get container status \"a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a\": rpc error: code = NotFound desc = could not find container \"a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a\": container with ID starting with a81799ec4ccf590b178269642ff6a06feda3493a4a2d32ca233bbb7f65e5ac5a not found: ID does not exist" Oct 04 04:59:40 crc kubenswrapper[4575]: I1004 04:59:40.036049 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:59:40 crc kubenswrapper[4575]: I1004 04:59:40.071888 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:59:40 crc kubenswrapper[4575]: I1004 04:59:40.079170 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b91f552-6eb5-479d-9a79-07a356d33ab0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:40 crc kubenswrapper[4575]: I1004 04:59:40.192417 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qj7qr"] Oct 04 04:59:40 crc kubenswrapper[4575]: I1004 04:59:40.205448 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qj7qr"] Oct 04 04:59:40 crc kubenswrapper[4575]: I1004 04:59:40.239962 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:40 crc kubenswrapper[4575]: I1004 04:59:40.565152 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtc62" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:40 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:40 crc kubenswrapper[4575]: > Oct 04 04:59:40 crc kubenswrapper[4575]: I1004 04:59:40.823424 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerStarted","Data":"85ce104c11b08014d5814f2a93a8c3a15104e41b8f2135f591d9c898ea45af7f"} Oct 04 04:59:40 crc kubenswrapper[4575]: I1004 04:59:40.874182 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:59:41 crc kubenswrapper[4575]: I1004 04:59:41.327296 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" path="/var/lib/kubelet/pods/2b91f552-6eb5-479d-9a79-07a356d33ab0/volumes" Oct 04 04:59:41 crc kubenswrapper[4575]: I1004 04:59:41.912940 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerStarted","Data":"8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667"} Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.612776 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.743168 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdpwq\" (UniqueName: \"kubernetes.io/projected/915e37b0-2385-42b9-b9fe-a4b1612e00d6-kube-api-access-fdpwq\") pod \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.743362 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-combined-ca-bundle\") pod \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.743498 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-config-data\") pod \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\" (UID: \"915e37b0-2385-42b9-b9fe-a4b1612e00d6\") " Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.753950 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/915e37b0-2385-42b9-b9fe-a4b1612e00d6-kube-api-access-fdpwq" (OuterVolumeSpecName: "kube-api-access-fdpwq") pod "915e37b0-2385-42b9-b9fe-a4b1612e00d6" (UID: "915e37b0-2385-42b9-b9fe-a4b1612e00d6"). InnerVolumeSpecName "kube-api-access-fdpwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.799402 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "915e37b0-2385-42b9-b9fe-a4b1612e00d6" (UID: "915e37b0-2385-42b9-b9fe-a4b1612e00d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.816858 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-config-data" (OuterVolumeSpecName: "config-data") pod "915e37b0-2385-42b9-b9fe-a4b1612e00d6" (UID: "915e37b0-2385-42b9-b9fe-a4b1612e00d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.846938 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdpwq\" (UniqueName: \"kubernetes.io/projected/915e37b0-2385-42b9-b9fe-a4b1612e00d6-kube-api-access-fdpwq\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.846969 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.846978 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/915e37b0-2385-42b9-b9fe-a4b1612e00d6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.927408 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerStarted","Data":"23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964"} Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.932152 4575 generic.go:334] "Generic (PLEG): container finished" podID="915e37b0-2385-42b9-b9fe-a4b1612e00d6" containerID="3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb" exitCode=137 Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.932205 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"915e37b0-2385-42b9-b9fe-a4b1612e00d6","Type":"ContainerDied","Data":"3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb"} Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.932213 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.932237 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"915e37b0-2385-42b9-b9fe-a4b1612e00d6","Type":"ContainerDied","Data":"95ea5541ca348f0d49fd94be947f9e0f6d29407a165418275c960be43871d5dc"} Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.932260 4575 scope.go:117] "RemoveContainer" containerID="3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.965404 4575 scope.go:117] "RemoveContainer" containerID="3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb" Oct 04 04:59:42 crc kubenswrapper[4575]: E1004 04:59:42.966188 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb\": container with ID starting with 3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb not found: ID does not exist" containerID="3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.966229 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb"} err="failed to get container status \"3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb\": rpc error: code = NotFound desc = could not find container \"3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb\": container with ID starting with 3b0314582611ce96aacaf10f19d23d1304a44dcd94fb99ffed131a334463aafb not found: ID does not exist" Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.986000 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:59:42 crc kubenswrapper[4575]: I1004 04:59:42.997166 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.015120 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.015723 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.026344 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:59:43 crc kubenswrapper[4575]: E1004 04:59:43.027146 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerName="extract-content" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.027264 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerName="extract-content" Oct 04 04:59:43 crc kubenswrapper[4575]: E1004 04:59:43.027370 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerName="extract-utilities" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.027454 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerName="extract-utilities" Oct 04 04:59:43 crc kubenswrapper[4575]: E1004 04:59:43.027784 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="915e37b0-2385-42b9-b9fe-a4b1612e00d6" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.027881 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="915e37b0-2385-42b9-b9fe-a4b1612e00d6" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:59:43 crc kubenswrapper[4575]: E1004 04:59:43.027983 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerName="registry-server" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.028071 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerName="registry-server" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.028461 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="915e37b0-2385-42b9-b9fe-a4b1612e00d6" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.028566 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b91f552-6eb5-479d-9a79-07a356d33ab0" containerName="registry-server" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.029540 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.035201 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.035362 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.035878 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.049228 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.152761 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmjks\" (UniqueName: \"kubernetes.io/projected/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-kube-api-access-rmjks\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.153301 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.153479 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.153649 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.154151 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.255857 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.255935 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.255967 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.256004 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.256073 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmjks\" (UniqueName: \"kubernetes.io/projected/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-kube-api-access-rmjks\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.265253 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.266301 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.273202 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.274177 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.304262 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmjks\" (UniqueName: \"kubernetes.io/projected/4ddc5e5c-23f0-4bcc-8f6b-32312096a689-kube-api-access-rmjks\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ddc5e5c-23f0-4bcc-8f6b-32312096a689\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.354012 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="915e37b0-2385-42b9-b9fe-a4b1612e00d6" path="/var/lib/kubelet/pods/915e37b0-2385-42b9-b9fe-a4b1612e00d6/volumes" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.364771 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.390932 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.504726 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.996405 4575 generic.go:334] "Generic (PLEG): container finished" podID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerID="96385a08097f46b617b586399b842fecc52bc222eec389ab74b37fba5f2df3e2" exitCode=137 Oct 04 04:59:43 crc kubenswrapper[4575]: I1004 04:59:43.996838 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18302821-ac2c-4ca2-9f33-86ba861a9e0a","Type":"ContainerDied","Data":"96385a08097f46b617b586399b842fecc52bc222eec389ab74b37fba5f2df3e2"} Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.034767 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.036580 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerStarted","Data":"09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af"} Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.062842 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.071631 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4ml2t"] Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.114655 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.392551 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.428962 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-combined-ca-bundle\") pod \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.429024 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-config-data\") pod \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.429112 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18302821-ac2c-4ca2-9f33-86ba861a9e0a-logs\") pod \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.429580 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpk84\" (UniqueName: \"kubernetes.io/projected/18302821-ac2c-4ca2-9f33-86ba861a9e0a-kube-api-access-cpk84\") pod \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\" (UID: \"18302821-ac2c-4ca2-9f33-86ba861a9e0a\") " Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.448204 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18302821-ac2c-4ca2-9f33-86ba861a9e0a-logs" (OuterVolumeSpecName: "logs") pod "18302821-ac2c-4ca2-9f33-86ba861a9e0a" (UID: "18302821-ac2c-4ca2-9f33-86ba861a9e0a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.454153 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18302821-ac2c-4ca2-9f33-86ba861a9e0a-kube-api-access-cpk84" (OuterVolumeSpecName: "kube-api-access-cpk84") pod "18302821-ac2c-4ca2-9f33-86ba861a9e0a" (UID: "18302821-ac2c-4ca2-9f33-86ba861a9e0a"). InnerVolumeSpecName "kube-api-access-cpk84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.505518 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18302821-ac2c-4ca2-9f33-86ba861a9e0a" (UID: "18302821-ac2c-4ca2-9f33-86ba861a9e0a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.540839 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.540875 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18302821-ac2c-4ca2-9f33-86ba861a9e0a-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.540890 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpk84\" (UniqueName: \"kubernetes.io/projected/18302821-ac2c-4ca2-9f33-86ba861a9e0a-kube-api-access-cpk84\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.574778 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-config-data" (OuterVolumeSpecName: "config-data") pod "18302821-ac2c-4ca2-9f33-86ba861a9e0a" (UID: "18302821-ac2c-4ca2-9f33-86ba861a9e0a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:44 crc kubenswrapper[4575]: I1004 04:59:44.642127 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18302821-ac2c-4ca2-9f33-86ba861a9e0a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.069773 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"18302821-ac2c-4ca2-9f33-86ba861a9e0a","Type":"ContainerDied","Data":"ab4fc1cfd385d3f21c1a6ba0cf2bb39099c474db7650db1e9f275ad4e10b6f57"} Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.070165 4575 scope.go:117] "RemoveContainer" containerID="96385a08097f46b617b586399b842fecc52bc222eec389ab74b37fba5f2df3e2" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.070407 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.074971 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerStarted","Data":"972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af"} Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.075145 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.079252 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ddc5e5c-23f0-4bcc-8f6b-32312096a689","Type":"ContainerStarted","Data":"4c8374b55c1dbcacabbc5836d41ca687a3a66e0c71ed951aa877bc7517923e97"} Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.079400 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4ml2t" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="registry-server" containerID="cri-o://eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028" gracePeriod=2 Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.079430 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ddc5e5c-23f0-4bcc-8f6b-32312096a689","Type":"ContainerStarted","Data":"bdd2022a0f8438d90fc0a662d71c2e5a1f7ee0c23cd5dc27549f53d0fcaee14c"} Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.110276 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.962167132 podStartE2EDuration="6.110251118s" podCreationTimestamp="2025-10-04 04:59:39 +0000 UTC" firstStartedPulling="2025-10-04 04:59:40.239487006 +0000 UTC m=+1531.568045810" lastFinishedPulling="2025-10-04 04:59:44.387570982 +0000 UTC m=+1535.716129796" observedRunningTime="2025-10-04 04:59:45.105405238 +0000 UTC m=+1536.433964052" watchObservedRunningTime="2025-10-04 04:59:45.110251118 +0000 UTC m=+1536.438809932" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.147039 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.147015219 podStartE2EDuration="3.147015219s" podCreationTimestamp="2025-10-04 04:59:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:45.145071143 +0000 UTC m=+1536.473629967" watchObservedRunningTime="2025-10-04 04:59:45.147015219 +0000 UTC m=+1536.475574033" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.168848 4575 scope.go:117] "RemoveContainer" containerID="3910011c01da877db2d770d227425f7450dff968195a5879ab0011eb13e09eda" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.179534 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.198943 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.228757 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:59:45 crc kubenswrapper[4575]: E1004 04:59:45.229293 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerName="nova-metadata-log" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.229312 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerName="nova-metadata-log" Oct 04 04:59:45 crc kubenswrapper[4575]: E1004 04:59:45.229345 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerName="nova-metadata-metadata" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.229353 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerName="nova-metadata-metadata" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.229616 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerName="nova-metadata-log" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.229633 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" containerName="nova-metadata-metadata" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.230752 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.237100 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.243980 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.240296 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.262362 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-config-data\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.262445 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrkk6\" (UniqueName: \"kubernetes.io/projected/a10047f1-cb48-4f95-a18b-2dd340669569-kube-api-access-xrkk6\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.262548 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.262578 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.262733 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10047f1-cb48-4f95-a18b-2dd340669569-logs\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.339217 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18302821-ac2c-4ca2-9f33-86ba861a9e0a" path="/var/lib/kubelet/pods/18302821-ac2c-4ca2-9f33-86ba861a9e0a/volumes" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.367068 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.367145 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.367248 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10047f1-cb48-4f95-a18b-2dd340669569-logs\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.367416 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-config-data\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.367476 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrkk6\" (UniqueName: \"kubernetes.io/projected/a10047f1-cb48-4f95-a18b-2dd340669569-kube-api-access-xrkk6\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.371461 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10047f1-cb48-4f95-a18b-2dd340669569-logs\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.380493 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.381469 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.382361 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-config-data\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.404465 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrkk6\" (UniqueName: \"kubernetes.io/projected/a10047f1-cb48-4f95-a18b-2dd340669569-kube-api-access-xrkk6\") pod \"nova-metadata-0\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.559278 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.666875 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.714035 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.714323 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.727443 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.915333 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.989404 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-catalog-content\") pod \"8da33bfd-1332-4c02-b37a-3605982815e8\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.989561 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-utilities\") pod \"8da33bfd-1332-4c02-b37a-3605982815e8\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.989728 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4dgp\" (UniqueName: \"kubernetes.io/projected/8da33bfd-1332-4c02-b37a-3605982815e8-kube-api-access-f4dgp\") pod \"8da33bfd-1332-4c02-b37a-3605982815e8\" (UID: \"8da33bfd-1332-4c02-b37a-3605982815e8\") " Oct 04 04:59:45 crc kubenswrapper[4575]: I1004 04:59:45.995310 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-utilities" (OuterVolumeSpecName: "utilities") pod "8da33bfd-1332-4c02-b37a-3605982815e8" (UID: "8da33bfd-1332-4c02-b37a-3605982815e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.038016 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8da33bfd-1332-4c02-b37a-3605982815e8-kube-api-access-f4dgp" (OuterVolumeSpecName: "kube-api-access-f4dgp") pod "8da33bfd-1332-4c02-b37a-3605982815e8" (UID: "8da33bfd-1332-4c02-b37a-3605982815e8"). InnerVolumeSpecName "kube-api-access-f4dgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.091901 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.091945 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4dgp\" (UniqueName: \"kubernetes.io/projected/8da33bfd-1332-4c02-b37a-3605982815e8-kube-api-access-f4dgp\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.119844 4575 generic.go:334] "Generic (PLEG): container finished" podID="8da33bfd-1332-4c02-b37a-3605982815e8" containerID="eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028" exitCode=0 Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.119940 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4ml2t" event={"ID":"8da33bfd-1332-4c02-b37a-3605982815e8","Type":"ContainerDied","Data":"eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028"} Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.119985 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4ml2t" event={"ID":"8da33bfd-1332-4c02-b37a-3605982815e8","Type":"ContainerDied","Data":"56c16780e51bdb2d665313fb5e573d2c575c5bb1d4a34699b5e8a262fbe51114"} Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.119983 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4ml2t" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.120104 4575 scope.go:117] "RemoveContainer" containerID="eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.157016 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8da33bfd-1332-4c02-b37a-3605982815e8" (UID: "8da33bfd-1332-4c02-b37a-3605982815e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.195834 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8da33bfd-1332-4c02-b37a-3605982815e8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.205979 4575 scope.go:117] "RemoveContainer" containerID="42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.254019 4575 scope.go:117] "RemoveContainer" containerID="c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.326501 4575 scope.go:117] "RemoveContainer" containerID="eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028" Oct 04 04:59:46 crc kubenswrapper[4575]: E1004 04:59:46.332055 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028\": container with ID starting with eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028 not found: ID does not exist" containerID="eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.332115 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028"} err="failed to get container status \"eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028\": rpc error: code = NotFound desc = could not find container \"eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028\": container with ID starting with eeba503fa53a765d97b137170dbffaade1f70ce873aa1f5d4f12118ff13c2028 not found: ID does not exist" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.332146 4575 scope.go:117] "RemoveContainer" containerID="42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af" Oct 04 04:59:46 crc kubenswrapper[4575]: E1004 04:59:46.332657 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af\": container with ID starting with 42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af not found: ID does not exist" containerID="42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.332691 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af"} err="failed to get container status \"42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af\": rpc error: code = NotFound desc = could not find container \"42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af\": container with ID starting with 42bc8abb8c7e0e6436969058101221dbe647b1c6212dc76406c033c0620ee2af not found: ID does not exist" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.332713 4575 scope.go:117] "RemoveContainer" containerID="c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244" Oct 04 04:59:46 crc kubenswrapper[4575]: E1004 04:59:46.333013 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244\": container with ID starting with c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244 not found: ID does not exist" containerID="c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.333040 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244"} err="failed to get container status \"c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244\": rpc error: code = NotFound desc = could not find container \"c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244\": container with ID starting with c3a3a18cdba6908200c0ebd23a85503b2e38424d876b485d8efa50723fac7244 not found: ID does not exist" Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.390660 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.476666 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4ml2t"] Oct 04 04:59:46 crc kubenswrapper[4575]: I1004 04:59:46.484992 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4ml2t"] Oct 04 04:59:47 crc kubenswrapper[4575]: I1004 04:59:47.176057 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a10047f1-cb48-4f95-a18b-2dd340669569","Type":"ContainerStarted","Data":"7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3"} Oct 04 04:59:47 crc kubenswrapper[4575]: I1004 04:59:47.176455 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a10047f1-cb48-4f95-a18b-2dd340669569","Type":"ContainerStarted","Data":"70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7"} Oct 04 04:59:47 crc kubenswrapper[4575]: I1004 04:59:47.176472 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a10047f1-cb48-4f95-a18b-2dd340669569","Type":"ContainerStarted","Data":"cf628e9ae6e08ed7de73939d19e485ebb3861e9ac56f0408de20b38c93799baa"} Oct 04 04:59:47 crc kubenswrapper[4575]: I1004 04:59:47.179337 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 04:59:47 crc kubenswrapper[4575]: I1004 04:59:47.204896 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.204878101 podStartE2EDuration="2.204878101s" podCreationTimestamp="2025-10-04 04:59:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:47.196133009 +0000 UTC m=+1538.524691823" watchObservedRunningTime="2025-10-04 04:59:47.204878101 +0000 UTC m=+1538.533436915" Oct 04 04:59:47 crc kubenswrapper[4575]: I1004 04:59:47.353276 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" path="/var/lib/kubelet/pods/8da33bfd-1332-4c02-b37a-3605982815e8/volumes" Oct 04 04:59:48 crc kubenswrapper[4575]: I1004 04:59:48.365538 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:50 crc kubenswrapper[4575]: I1004 04:59:50.512817 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtc62" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:50 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:50 crc kubenswrapper[4575]: > Oct 04 04:59:50 crc kubenswrapper[4575]: I1004 04:59:50.667705 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:59:50 crc kubenswrapper[4575]: I1004 04:59:50.668252 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.021127 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.022245 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.025083 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.030703 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.235420 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.238698 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.378082 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.421502 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.474371 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-cx2ll"] Oct 04 04:59:53 crc kubenswrapper[4575]: E1004 04:59:53.474916 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="registry-server" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.474938 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="registry-server" Oct 04 04:59:53 crc kubenswrapper[4575]: E1004 04:59:53.474963 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="extract-utilities" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.474971 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="extract-utilities" Oct 04 04:59:53 crc kubenswrapper[4575]: E1004 04:59:53.475002 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="extract-content" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.475009 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="extract-content" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.475234 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="8da33bfd-1332-4c02-b37a-3605982815e8" containerName="registry-server" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.476499 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.620855 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-cx2ll"] Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.697928 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlxql\" (UniqueName: \"kubernetes.io/projected/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-kube-api-access-zlxql\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.697998 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.698045 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.698180 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.698204 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.698256 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-config\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.800201 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.800261 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.800284 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.800336 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-config\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.800405 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlxql\" (UniqueName: \"kubernetes.io/projected/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-kube-api-access-zlxql\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.800438 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.801416 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.801558 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.801650 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.801835 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.802691 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-config\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:53 crc kubenswrapper[4575]: I1004 04:59:53.846961 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlxql\" (UniqueName: \"kubernetes.io/projected/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-kube-api-access-zlxql\") pod \"dnsmasq-dns-5c7b6c5df9-cx2ll\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.110133 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.304033 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.605702 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-k4drx"] Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.607313 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.612162 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.612367 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.621790 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-config-data\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.622217 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-scripts\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.622273 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l826\" (UniqueName: \"kubernetes.io/projected/7d863756-7f8d-4e1c-ad16-b44be09678c1-kube-api-access-5l826\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.622370 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.630788 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-k4drx"] Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.724328 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-config-data\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.724417 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-scripts\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.724472 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l826\" (UniqueName: \"kubernetes.io/projected/7d863756-7f8d-4e1c-ad16-b44be09678c1-kube-api-access-5l826\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.724570 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.735568 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-config-data\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.737062 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.745136 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-cx2ll"] Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.748395 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-scripts\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.780816 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l826\" (UniqueName: \"kubernetes.io/projected/7d863756-7f8d-4e1c-ad16-b44be09678c1-kube-api-access-5l826\") pod \"nova-cell1-cell-mapping-k4drx\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:54 crc kubenswrapper[4575]: I1004 04:59:54.939175 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 04:59:55 crc kubenswrapper[4575]: I1004 04:59:55.269806 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" event={"ID":"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443","Type":"ContainerStarted","Data":"e3bfef8b0c4bc46bbcc465711ecf12a02e3cb5445bd552bda3b91e5591a8fc86"} Oct 04 04:59:55 crc kubenswrapper[4575]: I1004 04:59:55.440756 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-k4drx"] Oct 04 04:59:55 crc kubenswrapper[4575]: I1004 04:59:55.557928 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 04:59:55 crc kubenswrapper[4575]: I1004 04:59:55.668491 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:59:55 crc kubenswrapper[4575]: I1004 04:59:55.669853 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:59:55 crc kubenswrapper[4575]: I1004 04:59:55.714434 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.281493 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-k4drx" event={"ID":"7d863756-7f8d-4e1c-ad16-b44be09678c1","Type":"ContainerStarted","Data":"dd570d023229065a4fc9fdafdb94922c41e48208913becd257a5357f7f92c95c"} Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.281982 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-k4drx" event={"ID":"7d863756-7f8d-4e1c-ad16-b44be09678c1","Type":"ContainerStarted","Data":"4b459abced3de7235fa899137a045e721a40bad6c5342d4d5cfc6ed5f660633f"} Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.284334 4575 generic.go:334] "Generic (PLEG): container finished" podID="ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" containerID="3e6c1087351b9cdd29bc0a5fea0e2388643d779e44f47149978f74bd53089d36" exitCode=0 Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.284964 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" event={"ID":"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443","Type":"ContainerDied","Data":"3e6c1087351b9cdd29bc0a5fea0e2388643d779e44f47149978f74bd53089d36"} Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.326534 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-k4drx" podStartSLOduration=2.326509695 podStartE2EDuration="2.326509695s" podCreationTimestamp="2025-10-04 04:59:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:56.316120665 +0000 UTC m=+1547.644679499" watchObservedRunningTime="2025-10-04 04:59:56.326509695 +0000 UTC m=+1547.655068509" Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.378251 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.378762 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="ceilometer-central-agent" containerID="cri-o://8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667" gracePeriod=30 Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.379503 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="proxy-httpd" containerID="cri-o://972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af" gracePeriod=30 Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.379554 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="sg-core" containerID="cri-o://09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af" gracePeriod=30 Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.383602 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="ceilometer-notification-agent" containerID="cri-o://23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964" gracePeriod=30 Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.405579 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.677739 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:59:56 crc kubenswrapper[4575]: I1004 04:59:56.677760 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.235225 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.235951 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-log" containerID="cri-o://72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a" gracePeriod=30 Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.235965 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-api" containerID="cri-o://2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4" gracePeriod=30 Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.347701 4575 generic.go:334] "Generic (PLEG): container finished" podID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerID="972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af" exitCode=0 Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.347728 4575 generic.go:334] "Generic (PLEG): container finished" podID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerID="09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af" exitCode=2 Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.347737 4575 generic.go:334] "Generic (PLEG): container finished" podID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerID="8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667" exitCode=0 Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.353794 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.353828 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" event={"ID":"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443","Type":"ContainerStarted","Data":"84f729f8e5b939d7e4745c7ca6e459581f6de204fe07d1cce6aed2995b8b7ba1"} Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.353844 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerDied","Data":"972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af"} Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.353857 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerDied","Data":"09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af"} Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.353866 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerDied","Data":"8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667"} Oct 04 04:59:57 crc kubenswrapper[4575]: I1004 04:59:57.388417 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" podStartSLOduration=4.388398407 podStartE2EDuration="4.388398407s" podCreationTimestamp="2025-10-04 04:59:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:59:57.375379371 +0000 UTC m=+1548.703938185" watchObservedRunningTime="2025-10-04 04:59:57.388398407 +0000 UTC m=+1548.716957221" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:57.927157 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.101947 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-ceilometer-tls-certs\") pod \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.101992 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-scripts\") pod \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.102025 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-log-httpd\") pod \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.102067 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-config-data\") pod \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.102089 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-run-httpd\") pod \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.102124 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6czkd\" (UniqueName: \"kubernetes.io/projected/5bc4c7a4-dd7f-4e32-aebe-79342e605977-kube-api-access-6czkd\") pod \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.102139 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-sg-core-conf-yaml\") pod \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.102166 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-combined-ca-bundle\") pod \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\" (UID: \"5bc4c7a4-dd7f-4e32-aebe-79342e605977\") " Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.105078 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5bc4c7a4-dd7f-4e32-aebe-79342e605977" (UID: "5bc4c7a4-dd7f-4e32-aebe-79342e605977"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.112666 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5bc4c7a4-dd7f-4e32-aebe-79342e605977" (UID: "5bc4c7a4-dd7f-4e32-aebe-79342e605977"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.117599 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc4c7a4-dd7f-4e32-aebe-79342e605977-kube-api-access-6czkd" (OuterVolumeSpecName: "kube-api-access-6czkd") pod "5bc4c7a4-dd7f-4e32-aebe-79342e605977" (UID: "5bc4c7a4-dd7f-4e32-aebe-79342e605977"). InnerVolumeSpecName "kube-api-access-6czkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.122696 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-scripts" (OuterVolumeSpecName: "scripts") pod "5bc4c7a4-dd7f-4e32-aebe-79342e605977" (UID: "5bc4c7a4-dd7f-4e32-aebe-79342e605977"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.197777 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5bc4c7a4-dd7f-4e32-aebe-79342e605977" (UID: "5bc4c7a4-dd7f-4e32-aebe-79342e605977"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.211052 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6czkd\" (UniqueName: \"kubernetes.io/projected/5bc4c7a4-dd7f-4e32-aebe-79342e605977-kube-api-access-6czkd\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.211087 4575 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.211099 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.211109 4575 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.211120 4575 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bc4c7a4-dd7f-4e32-aebe-79342e605977-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.262492 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5bc4c7a4-dd7f-4e32-aebe-79342e605977" (UID: "5bc4c7a4-dd7f-4e32-aebe-79342e605977"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.311997 4575 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.333705 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bc4c7a4-dd7f-4e32-aebe-79342e605977" (UID: "5bc4c7a4-dd7f-4e32-aebe-79342e605977"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.382784 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-config-data" (OuterVolumeSpecName: "config-data") pod "5bc4c7a4-dd7f-4e32-aebe-79342e605977" (UID: "5bc4c7a4-dd7f-4e32-aebe-79342e605977"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.383101 4575 generic.go:334] "Generic (PLEG): container finished" podID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerID="23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964" exitCode=0 Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.383189 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerDied","Data":"23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964"} Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.383217 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bc4c7a4-dd7f-4e32-aebe-79342e605977","Type":"ContainerDied","Data":"85ce104c11b08014d5814f2a93a8c3a15104e41b8f2135f591d9c898ea45af7f"} Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.383234 4575 scope.go:117] "RemoveContainer" containerID="972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.383253 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.386196 4575 generic.go:334] "Generic (PLEG): container finished" podID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerID="72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a" exitCode=143 Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.386748 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf8a470d-7d56-4511-b59a-3c4063676ba8","Type":"ContainerDied","Data":"72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a"} Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.420898 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.420933 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc4c7a4-dd7f-4e32-aebe-79342e605977-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.514500 4575 scope.go:117] "RemoveContainer" containerID="09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.534713 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.565662 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.588756 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:58 crc kubenswrapper[4575]: E1004 04:59:58.589189 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="proxy-httpd" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.589202 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="proxy-httpd" Oct 04 04:59:58 crc kubenswrapper[4575]: E1004 04:59:58.589219 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="ceilometer-central-agent" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.589226 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="ceilometer-central-agent" Oct 04 04:59:58 crc kubenswrapper[4575]: E1004 04:59:58.589282 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="ceilometer-notification-agent" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.589291 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="ceilometer-notification-agent" Oct 04 04:59:58 crc kubenswrapper[4575]: E1004 04:59:58.589315 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="sg-core" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.589321 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="sg-core" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.589523 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="ceilometer-central-agent" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.589542 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="ceilometer-notification-agent" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.589552 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="sg-core" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.589559 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" containerName="proxy-httpd" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.591215 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.583460 4575 scope.go:117] "RemoveContainer" containerID="23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.596236 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.596559 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.601037 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.623252 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.660772 4575 scope.go:117] "RemoveContainer" containerID="8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.710469 4575 scope.go:117] "RemoveContainer" containerID="972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af" Oct 04 04:59:58 crc kubenswrapper[4575]: E1004 04:59:58.710897 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af\": container with ID starting with 972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af not found: ID does not exist" containerID="972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.710948 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af"} err="failed to get container status \"972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af\": rpc error: code = NotFound desc = could not find container \"972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af\": container with ID starting with 972ce5944937067bbb697b63e464844a8739d9ee07e96114ba55a6a2c97a02af not found: ID does not exist" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.710966 4575 scope.go:117] "RemoveContainer" containerID="09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af" Oct 04 04:59:58 crc kubenswrapper[4575]: E1004 04:59:58.711188 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af\": container with ID starting with 09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af not found: ID does not exist" containerID="09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.711223 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af"} err="failed to get container status \"09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af\": rpc error: code = NotFound desc = could not find container \"09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af\": container with ID starting with 09d5cec869ad63deabdc484c10796ec528a318d29d91ef1726519fc419fcb2af not found: ID does not exist" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.711236 4575 scope.go:117] "RemoveContainer" containerID="23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964" Oct 04 04:59:58 crc kubenswrapper[4575]: E1004 04:59:58.711429 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964\": container with ID starting with 23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964 not found: ID does not exist" containerID="23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.711443 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964"} err="failed to get container status \"23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964\": rpc error: code = NotFound desc = could not find container \"23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964\": container with ID starting with 23e73f84f357191f661319f74701f2da830bedb223309f0cc7be2f321cc50964 not found: ID does not exist" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.711474 4575 scope.go:117] "RemoveContainer" containerID="8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667" Oct 04 04:59:58 crc kubenswrapper[4575]: E1004 04:59:58.711675 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667\": container with ID starting with 8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667 not found: ID does not exist" containerID="8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.711711 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667"} err="failed to get container status \"8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667\": rpc error: code = NotFound desc = could not find container \"8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667\": container with ID starting with 8ea72b16e18899acacbbb03dc655fb902d6f3340860ec45d6bc83bb10c30a667 not found: ID does not exist" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.730720 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.730787 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.731564 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-run-httpd\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.731652 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-scripts\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.731715 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.731741 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlk5r\" (UniqueName: \"kubernetes.io/projected/ed61f36b-b089-44aa-a413-616112487044-kube-api-access-qlk5r\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.731821 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-config-data\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.731958 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-log-httpd\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.833477 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.833517 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlk5r\" (UniqueName: \"kubernetes.io/projected/ed61f36b-b089-44aa-a413-616112487044-kube-api-access-qlk5r\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.833565 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-config-data\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.833662 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-log-httpd\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.833722 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.833750 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.833770 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-run-httpd\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.833804 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-scripts\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.835167 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-log-httpd\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.836145 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-run-httpd\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.841278 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.841545 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.841551 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-scripts\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.842019 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.846450 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-config-data\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.852068 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlk5r\" (UniqueName: \"kubernetes.io/projected/ed61f36b-b089-44aa-a413-616112487044-kube-api-access-qlk5r\") pod \"ceilometer-0\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " pod="openstack/ceilometer-0" Oct 04 04:59:58 crc kubenswrapper[4575]: I1004 04:59:58.968470 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:59:59 crc kubenswrapper[4575]: I1004 04:59:59.339994 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc4c7a4-dd7f-4e32-aebe-79342e605977" path="/var/lib/kubelet/pods/5bc4c7a4-dd7f-4e32-aebe-79342e605977/volumes" Oct 04 04:59:59 crc kubenswrapper[4575]: I1004 04:59:59.506380 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:59:59 crc kubenswrapper[4575]: W1004 04:59:59.531040 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded61f36b_b089_44aa_a413_616112487044.slice/crio-bdc12cd43323965f2ad17f69c07468a2be2ae4437b5b96eba93e9c397c7df1a7 WatchSource:0}: Error finding container bdc12cd43323965f2ad17f69c07468a2be2ae4437b5b96eba93e9c397c7df1a7: Status 404 returned error can't find the container with id bdc12cd43323965f2ad17f69c07468a2be2ae4437b5b96eba93e9c397c7df1a7 Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.150772 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8"] Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.152791 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.156304 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.158889 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8"] Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.161136 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.186215 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-secret-volume\") pod \"collect-profiles-29325900-fkqd8\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.186290 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-config-volume\") pod \"collect-profiles-29325900-fkqd8\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.186330 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cc9d\" (UniqueName: \"kubernetes.io/projected/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-kube-api-access-7cc9d\") pod \"collect-profiles-29325900-fkqd8\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.288032 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-secret-volume\") pod \"collect-profiles-29325900-fkqd8\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.288398 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-config-volume\") pod \"collect-profiles-29325900-fkqd8\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.288447 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cc9d\" (UniqueName: \"kubernetes.io/projected/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-kube-api-access-7cc9d\") pod \"collect-profiles-29325900-fkqd8\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.289949 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-config-volume\") pod \"collect-profiles-29325900-fkqd8\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.293940 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-secret-volume\") pod \"collect-profiles-29325900-fkqd8\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.314337 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cc9d\" (UniqueName: \"kubernetes.io/projected/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-kube-api-access-7cc9d\") pod \"collect-profiles-29325900-fkqd8\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.418813 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerStarted","Data":"bdc12cd43323965f2ad17f69c07468a2be2ae4437b5b96eba93e9c397c7df1a7"} Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.490557 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.508182 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtc62" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="registry-server" probeResult="failure" output=< Oct 04 05:00:00 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:00:00 crc kubenswrapper[4575]: > Oct 04 05:00:00 crc kubenswrapper[4575]: I1004 05:00:00.683991 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.101572 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8"] Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.409335 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.449188 4575 generic.go:334] "Generic (PLEG): container finished" podID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerID="2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4" exitCode=0 Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.449399 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.449625 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf8a470d-7d56-4511-b59a-3c4063676ba8","Type":"ContainerDied","Data":"2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4"} Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.449659 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf8a470d-7d56-4511-b59a-3c4063676ba8","Type":"ContainerDied","Data":"32596259c505ab716d5cd86f58654c90bd0f5bf9e3b83ccd4758d10809202430"} Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.449676 4575 scope.go:117] "RemoveContainer" containerID="2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.470270 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerStarted","Data":"f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8"} Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.474734 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" event={"ID":"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec","Type":"ContainerStarted","Data":"65248bcf134347aa5d6096d9620651815b42c1c13aab81296a436a35ee6b9456"} Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.521356 4575 scope.go:117] "RemoveContainer" containerID="72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.530270 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kw2v\" (UniqueName: \"kubernetes.io/projected/cf8a470d-7d56-4511-b59a-3c4063676ba8-kube-api-access-5kw2v\") pod \"cf8a470d-7d56-4511-b59a-3c4063676ba8\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.530364 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-config-data\") pod \"cf8a470d-7d56-4511-b59a-3c4063676ba8\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.530438 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-combined-ca-bundle\") pod \"cf8a470d-7d56-4511-b59a-3c4063676ba8\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.530508 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8a470d-7d56-4511-b59a-3c4063676ba8-logs\") pod \"cf8a470d-7d56-4511-b59a-3c4063676ba8\" (UID: \"cf8a470d-7d56-4511-b59a-3c4063676ba8\") " Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.531501 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf8a470d-7d56-4511-b59a-3c4063676ba8-logs" (OuterVolumeSpecName: "logs") pod "cf8a470d-7d56-4511-b59a-3c4063676ba8" (UID: "cf8a470d-7d56-4511-b59a-3c4063676ba8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.536792 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf8a470d-7d56-4511-b59a-3c4063676ba8-kube-api-access-5kw2v" (OuterVolumeSpecName: "kube-api-access-5kw2v") pod "cf8a470d-7d56-4511-b59a-3c4063676ba8" (UID: "cf8a470d-7d56-4511-b59a-3c4063676ba8"). InnerVolumeSpecName "kube-api-access-5kw2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.568153 4575 scope.go:117] "RemoveContainer" containerID="2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4" Oct 04 05:00:01 crc kubenswrapper[4575]: E1004 05:00:01.573106 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4\": container with ID starting with 2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4 not found: ID does not exist" containerID="2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.573151 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4"} err="failed to get container status \"2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4\": rpc error: code = NotFound desc = could not find container \"2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4\": container with ID starting with 2e2f6a2e6f2c6c09ea436dfcd3d31385d612bdbc0d623d497a395475d28168c4 not found: ID does not exist" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.573178 4575 scope.go:117] "RemoveContainer" containerID="72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a" Oct 04 05:00:01 crc kubenswrapper[4575]: E1004 05:00:01.576946 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a\": container with ID starting with 72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a not found: ID does not exist" containerID="72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.576998 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a"} err="failed to get container status \"72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a\": rpc error: code = NotFound desc = could not find container \"72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a\": container with ID starting with 72fa626a466380b287c368a0de83709ccbece910f06886425737c34b41d6af1a not found: ID does not exist" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.580448 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-config-data" (OuterVolumeSpecName: "config-data") pod "cf8a470d-7d56-4511-b59a-3c4063676ba8" (UID: "cf8a470d-7d56-4511-b59a-3c4063676ba8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.606774 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf8a470d-7d56-4511-b59a-3c4063676ba8" (UID: "cf8a470d-7d56-4511-b59a-3c4063676ba8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.632903 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kw2v\" (UniqueName: \"kubernetes.io/projected/cf8a470d-7d56-4511-b59a-3c4063676ba8-kube-api-access-5kw2v\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.632942 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.632965 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8a470d-7d56-4511-b59a-3c4063676ba8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.632976 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8a470d-7d56-4511-b59a-3c4063676ba8-logs\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.798675 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.809657 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.814916 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:01 crc kubenswrapper[4575]: E1004 05:00:01.815308 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-api" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.815325 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-api" Oct 04 05:00:01 crc kubenswrapper[4575]: E1004 05:00:01.815343 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-log" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.815349 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-log" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.815544 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-api" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.815572 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" containerName="nova-api-log" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.816548 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.821079 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.821292 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.821512 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.881663 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.939711 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.939814 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.939847 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-config-data\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.939898 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcf3c083-541f-4717-9fb1-620a4efb749f-logs\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.939948 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fskbg\" (UniqueName: \"kubernetes.io/projected/bcf3c083-541f-4717-9fb1-620a4efb749f-kube-api-access-fskbg\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:01 crc kubenswrapper[4575]: I1004 05:00:01.939972 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-public-tls-certs\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.047836 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.048221 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-config-data\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.048300 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcf3c083-541f-4717-9fb1-620a4efb749f-logs\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.048344 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fskbg\" (UniqueName: \"kubernetes.io/projected/bcf3c083-541f-4717-9fb1-620a4efb749f-kube-api-access-fskbg\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.048740 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcf3c083-541f-4717-9fb1-620a4efb749f-logs\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.048837 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-public-tls-certs\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.049225 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.054372 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-public-tls-certs\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.057353 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-config-data\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.057800 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.061442 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.073453 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fskbg\" (UniqueName: \"kubernetes.io/projected/bcf3c083-541f-4717-9fb1-620a4efb749f-kube-api-access-fskbg\") pod \"nova-api-0\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.140118 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.496306 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerStarted","Data":"2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e"} Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.498539 4575 generic.go:334] "Generic (PLEG): container finished" podID="b9b88d73-0f40-457f-8ece-5cdb2b87a0ec" containerID="8dc871af7f20ebe706d30ede36fcfbc471d80d3cd68dbdc0ba13457f4c0b631f" exitCode=0 Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.498599 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" event={"ID":"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec","Type":"ContainerDied","Data":"8dc871af7f20ebe706d30ede36fcfbc471d80d3cd68dbdc0ba13457f4c0b631f"} Oct 04 05:00:02 crc kubenswrapper[4575]: I1004 05:00:02.743698 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:03 crc kubenswrapper[4575]: I1004 05:00:03.322289 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf8a470d-7d56-4511-b59a-3c4063676ba8" path="/var/lib/kubelet/pods/cf8a470d-7d56-4511-b59a-3c4063676ba8/volumes" Oct 04 05:00:03 crc kubenswrapper[4575]: I1004 05:00:03.510670 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bcf3c083-541f-4717-9fb1-620a4efb749f","Type":"ContainerStarted","Data":"d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208"} Oct 04 05:00:03 crc kubenswrapper[4575]: I1004 05:00:03.512014 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bcf3c083-541f-4717-9fb1-620a4efb749f","Type":"ContainerStarted","Data":"52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16"} Oct 04 05:00:03 crc kubenswrapper[4575]: I1004 05:00:03.512096 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bcf3c083-541f-4717-9fb1-620a4efb749f","Type":"ContainerStarted","Data":"cf386755012d7ed4c05579f1209b6ab98052199c6e492e5b4aff8b05aacc89b6"} Oct 04 05:00:03 crc kubenswrapper[4575]: I1004 05:00:03.518824 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerStarted","Data":"5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202"} Oct 04 05:00:03 crc kubenswrapper[4575]: I1004 05:00:03.536090 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.536070251 podStartE2EDuration="2.536070251s" podCreationTimestamp="2025-10-04 05:00:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:00:03.528438089 +0000 UTC m=+1554.856996913" watchObservedRunningTime="2025-10-04 05:00:03.536070251 +0000 UTC m=+1554.864629085" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.014786 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.113747 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.214255 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cc9d\" (UniqueName: \"kubernetes.io/projected/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-kube-api-access-7cc9d\") pod \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.214558 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-secret-volume\") pod \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.214704 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-config-volume\") pod \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\" (UID: \"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec\") " Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.216049 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-config-volume" (OuterVolumeSpecName: "config-volume") pod "b9b88d73-0f40-457f-8ece-5cdb2b87a0ec" (UID: "b9b88d73-0f40-457f-8ece-5cdb2b87a0ec"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.231145 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-kube-api-access-7cc9d" (OuterVolumeSpecName: "kube-api-access-7cc9d") pod "b9b88d73-0f40-457f-8ece-5cdb2b87a0ec" (UID: "b9b88d73-0f40-457f-8ece-5cdb2b87a0ec"). InnerVolumeSpecName "kube-api-access-7cc9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.244440 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b9b88d73-0f40-457f-8ece-5cdb2b87a0ec" (UID: "b9b88d73-0f40-457f-8ece-5cdb2b87a0ec"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.317249 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cc9d\" (UniqueName: \"kubernetes.io/projected/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-kube-api-access-7cc9d\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.317293 4575 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.317304 4575 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.381683 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vn9rj"] Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.381925 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" podUID="854288d3-ecc2-482f-a986-f27cc7672efa" containerName="dnsmasq-dns" containerID="cri-o://f4bb2f72787716c30050fdbe6afcc0302e5cd9b363fd6387254b80f9896ff6bd" gracePeriod=10 Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.530804 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" event={"ID":"b9b88d73-0f40-457f-8ece-5cdb2b87a0ec","Type":"ContainerDied","Data":"65248bcf134347aa5d6096d9620651815b42c1c13aab81296a436a35ee6b9456"} Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.530843 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65248bcf134347aa5d6096d9620651815b42c1c13aab81296a436a35ee6b9456" Oct 04 05:00:04 crc kubenswrapper[4575]: I1004 05:00:04.530841 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.567427 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.567859 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.569266 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"1286c3d99af99d1bd1f66544432932ee55a667d406d520ce4179dfad08a0c438"} pod="openstack/horizon-b86b879b4-tlf42" containerMessage="Container horizon failed startup probe, will be restarted" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.569308 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" containerID="cri-o://1286c3d99af99d1bd1f66544432932ee55a667d406d520ce4179dfad08a0c438" gracePeriod=30 Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.608539 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="ceilometer-central-agent" containerID="cri-o://f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8" gracePeriod=30 Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.608739 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerStarted","Data":"5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04"} Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.609363 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.608928 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="sg-core" containerID="cri-o://5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202" gracePeriod=30 Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.608941 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="ceilometer-notification-agent" containerID="cri-o://2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e" gracePeriod=30 Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.608899 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="proxy-httpd" containerID="cri-o://5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04" gracePeriod=30 Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.614838 4575 generic.go:334] "Generic (PLEG): container finished" podID="854288d3-ecc2-482f-a986-f27cc7672efa" containerID="f4bb2f72787716c30050fdbe6afcc0302e5cd9b363fd6387254b80f9896ff6bd" exitCode=0 Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.614886 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" event={"ID":"854288d3-ecc2-482f-a986-f27cc7672efa","Type":"ContainerDied","Data":"f4bb2f72787716c30050fdbe6afcc0302e5cd9b363fd6387254b80f9896ff6bd"} Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.614912 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" event={"ID":"854288d3-ecc2-482f-a986-f27cc7672efa","Type":"ContainerDied","Data":"2d538297cbfbed345effb9766b528351743d21527348d089232432a7faad8a04"} Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.614925 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d538297cbfbed345effb9766b528351743d21527348d089232432a7faad8a04" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.663357 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.665621 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.166918736 podStartE2EDuration="7.665605085s" podCreationTimestamp="2025-10-04 04:59:58 +0000 UTC" firstStartedPulling="2025-10-04 04:59:59.541570167 +0000 UTC m=+1550.870128981" lastFinishedPulling="2025-10-04 05:00:04.040256506 +0000 UTC m=+1555.368815330" observedRunningTime="2025-10-04 05:00:05.646186211 +0000 UTC m=+1556.974745035" watchObservedRunningTime="2025-10-04 05:00:05.665605085 +0000 UTC m=+1556.994163899" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.714018 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.714737 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.714792 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.715758 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"f8d4443c36b0c2245629253c64c8178d2ca6381e985a3cf457f5457982a4dcd8"} pod="openstack/horizon-644bf5cdd4-pwdw9" containerMessage="Container horizon failed startup probe, will be restarted" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.715800 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" containerID="cri-o://f8d4443c36b0c2245629253c64c8178d2ca6381e985a3cf457f5457982a4dcd8" gracePeriod=30 Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.721306 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.766476 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-config\") pod \"854288d3-ecc2-482f-a986-f27cc7672efa\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.766600 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-svc\") pod \"854288d3-ecc2-482f-a986-f27cc7672efa\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.766748 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-swift-storage-0\") pod \"854288d3-ecc2-482f-a986-f27cc7672efa\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.766797 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-sb\") pod \"854288d3-ecc2-482f-a986-f27cc7672efa\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.766881 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bjxg\" (UniqueName: \"kubernetes.io/projected/854288d3-ecc2-482f-a986-f27cc7672efa-kube-api-access-4bjxg\") pod \"854288d3-ecc2-482f-a986-f27cc7672efa\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.766914 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-nb\") pod \"854288d3-ecc2-482f-a986-f27cc7672efa\" (UID: \"854288d3-ecc2-482f-a986-f27cc7672efa\") " Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.791041 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.792824 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/854288d3-ecc2-482f-a986-f27cc7672efa-kube-api-access-4bjxg" (OuterVolumeSpecName: "kube-api-access-4bjxg") pod "854288d3-ecc2-482f-a986-f27cc7672efa" (UID: "854288d3-ecc2-482f-a986-f27cc7672efa"). InnerVolumeSpecName "kube-api-access-4bjxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.870782 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bjxg\" (UniqueName: \"kubernetes.io/projected/854288d3-ecc2-482f-a986-f27cc7672efa-kube-api-access-4bjxg\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.909860 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "854288d3-ecc2-482f-a986-f27cc7672efa" (UID: "854288d3-ecc2-482f-a986-f27cc7672efa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.911611 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-config" (OuterVolumeSpecName: "config") pod "854288d3-ecc2-482f-a986-f27cc7672efa" (UID: "854288d3-ecc2-482f-a986-f27cc7672efa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.914183 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "854288d3-ecc2-482f-a986-f27cc7672efa" (UID: "854288d3-ecc2-482f-a986-f27cc7672efa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.925999 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "854288d3-ecc2-482f-a986-f27cc7672efa" (UID: "854288d3-ecc2-482f-a986-f27cc7672efa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.942814 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "854288d3-ecc2-482f-a986-f27cc7672efa" (UID: "854288d3-ecc2-482f-a986-f27cc7672efa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.972513 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.972550 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.972562 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.972579 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-config\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:05 crc kubenswrapper[4575]: I1004 05:00:05.972705 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854288d3-ecc2-482f-a986-f27cc7672efa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.626534 4575 generic.go:334] "Generic (PLEG): container finished" podID="7d863756-7f8d-4e1c-ad16-b44be09678c1" containerID="dd570d023229065a4fc9fdafdb94922c41e48208913becd257a5357f7f92c95c" exitCode=0 Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.626670 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-k4drx" event={"ID":"7d863756-7f8d-4e1c-ad16-b44be09678c1","Type":"ContainerDied","Data":"dd570d023229065a4fc9fdafdb94922c41e48208913becd257a5357f7f92c95c"} Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.632414 4575 generic.go:334] "Generic (PLEG): container finished" podID="ed61f36b-b089-44aa-a413-616112487044" containerID="5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04" exitCode=0 Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.632452 4575 generic.go:334] "Generic (PLEG): container finished" podID="ed61f36b-b089-44aa-a413-616112487044" containerID="5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202" exitCode=2 Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.632463 4575 generic.go:334] "Generic (PLEG): container finished" podID="ed61f36b-b089-44aa-a413-616112487044" containerID="2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e" exitCode=0 Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.632542 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-vn9rj" Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.633754 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerDied","Data":"5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04"} Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.633815 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerDied","Data":"5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202"} Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.633831 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerDied","Data":"2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e"} Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.642399 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.705786 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vn9rj"] Oct 04 05:00:06 crc kubenswrapper[4575]: I1004 05:00:06.740174 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vn9rj"] Oct 04 05:00:07 crc kubenswrapper[4575]: I1004 05:00:07.323208 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="854288d3-ecc2-482f-a986-f27cc7672efa" path="/var/lib/kubelet/pods/854288d3-ecc2-482f-a986-f27cc7672efa/volumes" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.030185 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.111131 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-scripts\") pod \"7d863756-7f8d-4e1c-ad16-b44be09678c1\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.111192 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-combined-ca-bundle\") pod \"7d863756-7f8d-4e1c-ad16-b44be09678c1\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.111251 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-config-data\") pod \"7d863756-7f8d-4e1c-ad16-b44be09678c1\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.111354 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l826\" (UniqueName: \"kubernetes.io/projected/7d863756-7f8d-4e1c-ad16-b44be09678c1-kube-api-access-5l826\") pod \"7d863756-7f8d-4e1c-ad16-b44be09678c1\" (UID: \"7d863756-7f8d-4e1c-ad16-b44be09678c1\") " Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.117106 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d863756-7f8d-4e1c-ad16-b44be09678c1-kube-api-access-5l826" (OuterVolumeSpecName: "kube-api-access-5l826") pod "7d863756-7f8d-4e1c-ad16-b44be09678c1" (UID: "7d863756-7f8d-4e1c-ad16-b44be09678c1"). InnerVolumeSpecName "kube-api-access-5l826". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.118217 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-scripts" (OuterVolumeSpecName: "scripts") pod "7d863756-7f8d-4e1c-ad16-b44be09678c1" (UID: "7d863756-7f8d-4e1c-ad16-b44be09678c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.140014 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d863756-7f8d-4e1c-ad16-b44be09678c1" (UID: "7d863756-7f8d-4e1c-ad16-b44be09678c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.150708 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-config-data" (OuterVolumeSpecName: "config-data") pod "7d863756-7f8d-4e1c-ad16-b44be09678c1" (UID: "7d863756-7f8d-4e1c-ad16-b44be09678c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.213741 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.213770 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.213780 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d863756-7f8d-4e1c-ad16-b44be09678c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.213789 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l826\" (UniqueName: \"kubernetes.io/projected/7d863756-7f8d-4e1c-ad16-b44be09678c1-kube-api-access-5l826\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.446499 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.446562 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.654357 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-k4drx" event={"ID":"7d863756-7f8d-4e1c-ad16-b44be09678c1","Type":"ContainerDied","Data":"4b459abced3de7235fa899137a045e721a40bad6c5342d4d5cfc6ed5f660633f"} Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.654844 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b459abced3de7235fa899137a045e721a40bad6c5342d4d5cfc6ed5f660633f" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.654396 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-k4drx" Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.861603 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.861947 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerName="nova-api-log" containerID="cri-o://52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16" gracePeriod=30 Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.862494 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerName="nova-api-api" containerID="cri-o://d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208" gracePeriod=30 Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.888537 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.888926 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="15a3821f-f7ca-45b4-b667-d5985bc25f3a" containerName="nova-scheduler-scheduler" containerID="cri-o://c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c" gracePeriod=30 Oct 04 05:00:08 crc kubenswrapper[4575]: I1004 05:00:08.919077 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.524394 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.531533 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.569739 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.598220 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644561 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-combined-ca-bundle\") pod \"ed61f36b-b089-44aa-a413-616112487044\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644640 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-config-data\") pod \"ed61f36b-b089-44aa-a413-616112487044\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644712 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-scripts\") pod \"ed61f36b-b089-44aa-a413-616112487044\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644743 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-sg-core-conf-yaml\") pod \"ed61f36b-b089-44aa-a413-616112487044\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644832 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fskbg\" (UniqueName: \"kubernetes.io/projected/bcf3c083-541f-4717-9fb1-620a4efb749f-kube-api-access-fskbg\") pod \"bcf3c083-541f-4717-9fb1-620a4efb749f\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644869 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-internal-tls-certs\") pod \"bcf3c083-541f-4717-9fb1-620a4efb749f\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644897 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlk5r\" (UniqueName: \"kubernetes.io/projected/ed61f36b-b089-44aa-a413-616112487044-kube-api-access-qlk5r\") pod \"ed61f36b-b089-44aa-a413-616112487044\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644947 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcf3c083-541f-4717-9fb1-620a4efb749f-logs\") pod \"bcf3c083-541f-4717-9fb1-620a4efb749f\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644972 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-public-tls-certs\") pod \"bcf3c083-541f-4717-9fb1-620a4efb749f\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.644996 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-run-httpd\") pod \"ed61f36b-b089-44aa-a413-616112487044\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.645022 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-config-data\") pod \"bcf3c083-541f-4717-9fb1-620a4efb749f\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.645055 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-log-httpd\") pod \"ed61f36b-b089-44aa-a413-616112487044\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.645103 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-ceilometer-tls-certs\") pod \"ed61f36b-b089-44aa-a413-616112487044\" (UID: \"ed61f36b-b089-44aa-a413-616112487044\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.645128 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-combined-ca-bundle\") pod \"bcf3c083-541f-4717-9fb1-620a4efb749f\" (UID: \"bcf3c083-541f-4717-9fb1-620a4efb749f\") " Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.645551 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcf3c083-541f-4717-9fb1-620a4efb749f-logs" (OuterVolumeSpecName: "logs") pod "bcf3c083-541f-4717-9fb1-620a4efb749f" (UID: "bcf3c083-541f-4717-9fb1-620a4efb749f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.645932 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcf3c083-541f-4717-9fb1-620a4efb749f-logs\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.646711 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ed61f36b-b089-44aa-a413-616112487044" (UID: "ed61f36b-b089-44aa-a413-616112487044"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.647034 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ed61f36b-b089-44aa-a413-616112487044" (UID: "ed61f36b-b089-44aa-a413-616112487044"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.654239 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-scripts" (OuterVolumeSpecName: "scripts") pod "ed61f36b-b089-44aa-a413-616112487044" (UID: "ed61f36b-b089-44aa-a413-616112487044"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.654521 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcf3c083-541f-4717-9fb1-620a4efb749f-kube-api-access-fskbg" (OuterVolumeSpecName: "kube-api-access-fskbg") pod "bcf3c083-541f-4717-9fb1-620a4efb749f" (UID: "bcf3c083-541f-4717-9fb1-620a4efb749f"). InnerVolumeSpecName "kube-api-access-fskbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.657567 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed61f36b-b089-44aa-a413-616112487044-kube-api-access-qlk5r" (OuterVolumeSpecName: "kube-api-access-qlk5r") pod "ed61f36b-b089-44aa-a413-616112487044" (UID: "ed61f36b-b089-44aa-a413-616112487044"). InnerVolumeSpecName "kube-api-access-qlk5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.675622 4575 generic.go:334] "Generic (PLEG): container finished" podID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerID="d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208" exitCode=0 Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.675657 4575 generic.go:334] "Generic (PLEG): container finished" podID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerID="52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16" exitCode=143 Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.675726 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bcf3c083-541f-4717-9fb1-620a4efb749f","Type":"ContainerDied","Data":"d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208"} Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.675763 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bcf3c083-541f-4717-9fb1-620a4efb749f","Type":"ContainerDied","Data":"52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16"} Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.675776 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bcf3c083-541f-4717-9fb1-620a4efb749f","Type":"ContainerDied","Data":"cf386755012d7ed4c05579f1209b6ab98052199c6e492e5b4aff8b05aacc89b6"} Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.675793 4575 scope.go:117] "RemoveContainer" containerID="d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.675963 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.684861 4575 generic.go:334] "Generic (PLEG): container finished" podID="ed61f36b-b089-44aa-a413-616112487044" containerID="f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8" exitCode=0 Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.685056 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerDied","Data":"f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8"} Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.685121 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed61f36b-b089-44aa-a413-616112487044","Type":"ContainerDied","Data":"bdc12cd43323965f2ad17f69c07468a2be2ae4437b5b96eba93e9c397c7df1a7"} Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.685120 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-log" containerID="cri-o://70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7" gracePeriod=30 Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.685212 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-metadata" containerID="cri-o://7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3" gracePeriod=30 Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.685747 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.718610 4575 scope.go:117] "RemoveContainer" containerID="52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.724805 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-config-data" (OuterVolumeSpecName: "config-data") pod "bcf3c083-541f-4717-9fb1-620a4efb749f" (UID: "bcf3c083-541f-4717-9fb1-620a4efb749f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.747358 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.747390 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fskbg\" (UniqueName: \"kubernetes.io/projected/bcf3c083-541f-4717-9fb1-620a4efb749f-kube-api-access-fskbg\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.747402 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlk5r\" (UniqueName: \"kubernetes.io/projected/ed61f36b-b089-44aa-a413-616112487044-kube-api-access-qlk5r\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.747413 4575 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.747425 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.747440 4575 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed61f36b-b089-44aa-a413-616112487044-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.762447 4575 scope.go:117] "RemoveContainer" containerID="d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.762786 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ed61f36b-b089-44aa-a413-616112487044" (UID: "ed61f36b-b089-44aa-a413-616112487044"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: E1004 05:00:09.763361 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208\": container with ID starting with d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208 not found: ID does not exist" containerID="d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.763393 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208"} err="failed to get container status \"d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208\": rpc error: code = NotFound desc = could not find container \"d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208\": container with ID starting with d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208 not found: ID does not exist" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.763411 4575 scope.go:117] "RemoveContainer" containerID="52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16" Oct 04 05:00:09 crc kubenswrapper[4575]: E1004 05:00:09.763854 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16\": container with ID starting with 52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16 not found: ID does not exist" containerID="52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.763888 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16"} err="failed to get container status \"52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16\": rpc error: code = NotFound desc = could not find container \"52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16\": container with ID starting with 52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16 not found: ID does not exist" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.763917 4575 scope.go:117] "RemoveContainer" containerID="d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.763914 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcf3c083-541f-4717-9fb1-620a4efb749f" (UID: "bcf3c083-541f-4717-9fb1-620a4efb749f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.764286 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208"} err="failed to get container status \"d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208\": rpc error: code = NotFound desc = could not find container \"d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208\": container with ID starting with d9a12a29720ff12b893cf8f6becb96263db098c278e7a85f711003feae840208 not found: ID does not exist" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.764316 4575 scope.go:117] "RemoveContainer" containerID="52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.764566 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16"} err="failed to get container status \"52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16\": rpc error: code = NotFound desc = could not find container \"52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16\": container with ID starting with 52d66e818aa37aee5e3854234e60aab391831c6fc970f06014128ec1a0040d16 not found: ID does not exist" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.768676 4575 scope.go:117] "RemoveContainer" containerID="5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.773778 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dtc62"] Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.804485 4575 scope.go:117] "RemoveContainer" containerID="5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.811234 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bcf3c083-541f-4717-9fb1-620a4efb749f" (UID: "bcf3c083-541f-4717-9fb1-620a4efb749f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.820044 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ed61f36b-b089-44aa-a413-616112487044" (UID: "ed61f36b-b089-44aa-a413-616112487044"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.838802 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed61f36b-b089-44aa-a413-616112487044" (UID: "ed61f36b-b089-44aa-a413-616112487044"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.846224 4575 scope.go:117] "RemoveContainer" containerID="2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.852495 4575 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.852526 4575 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.852537 4575 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.852548 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.852559 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.853558 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bcf3c083-541f-4717-9fb1-620a4efb749f" (UID: "bcf3c083-541f-4717-9fb1-620a4efb749f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.901391 4575 scope.go:117] "RemoveContainer" containerID="f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.915433 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-config-data" (OuterVolumeSpecName: "config-data") pod "ed61f36b-b089-44aa-a413-616112487044" (UID: "ed61f36b-b089-44aa-a413-616112487044"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.923942 4575 scope.go:117] "RemoveContainer" containerID="5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04" Oct 04 05:00:09 crc kubenswrapper[4575]: E1004 05:00:09.924348 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04\": container with ID starting with 5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04 not found: ID does not exist" containerID="5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.924373 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04"} err="failed to get container status \"5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04\": rpc error: code = NotFound desc = could not find container \"5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04\": container with ID starting with 5eece9ebd498a0e9fe42c034e82f7e95a34c8dd6ef39b71a0658ebcdd465ee04 not found: ID does not exist" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.924391 4575 scope.go:117] "RemoveContainer" containerID="5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202" Oct 04 05:00:09 crc kubenswrapper[4575]: E1004 05:00:09.924769 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202\": container with ID starting with 5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202 not found: ID does not exist" containerID="5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.924787 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202"} err="failed to get container status \"5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202\": rpc error: code = NotFound desc = could not find container \"5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202\": container with ID starting with 5bbf5a6592977e28b8fe75b22541f2e209174d6806b7b6163458d30bb4304202 not found: ID does not exist" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.924799 4575 scope.go:117] "RemoveContainer" containerID="2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e" Oct 04 05:00:09 crc kubenswrapper[4575]: E1004 05:00:09.925242 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e\": container with ID starting with 2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e not found: ID does not exist" containerID="2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.925282 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e"} err="failed to get container status \"2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e\": rpc error: code = NotFound desc = could not find container \"2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e\": container with ID starting with 2dbaf53796fc82c685f6c8e4ede100ceece4fd53162ad0e463f533981569047e not found: ID does not exist" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.925307 4575 scope.go:117] "RemoveContainer" containerID="f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8" Oct 04 05:00:09 crc kubenswrapper[4575]: E1004 05:00:09.925606 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8\": container with ID starting with f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8 not found: ID does not exist" containerID="f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.925625 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8"} err="failed to get container status \"f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8\": rpc error: code = NotFound desc = could not find container \"f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8\": container with ID starting with f03cbabff25220a4073c66dc3b4526470a8471cb595b96d4afd4ee0044dcc9e8 not found: ID does not exist" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.954791 4575 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf3c083-541f-4717-9fb1-620a4efb749f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:09 crc kubenswrapper[4575]: I1004 05:00:09.954825 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed61f36b-b089-44aa-a413-616112487044-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.038884 4575 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.040768 4575 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.045191 4575 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.045272 4575 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="15a3821f-f7ca-45b4-b667-d5985bc25f3a" containerName="nova-scheduler-scheduler" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.088257 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.104806 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.120942 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.134930 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150087 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150604 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d863756-7f8d-4e1c-ad16-b44be09678c1" containerName="nova-manage" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150626 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d863756-7f8d-4e1c-ad16-b44be09678c1" containerName="nova-manage" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150637 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="ceilometer-notification-agent" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150645 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="ceilometer-notification-agent" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150660 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="proxy-httpd" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150668 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="proxy-httpd" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150680 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerName="nova-api-api" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150687 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerName="nova-api-api" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150710 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="ceilometer-central-agent" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150717 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="ceilometer-central-agent" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150726 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerName="nova-api-log" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150734 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerName="nova-api-log" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150742 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="sg-core" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150752 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="sg-core" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150776 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="854288d3-ecc2-482f-a986-f27cc7672efa" containerName="init" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150784 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="854288d3-ecc2-482f-a986-f27cc7672efa" containerName="init" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150808 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b88d73-0f40-457f-8ece-5cdb2b87a0ec" containerName="collect-profiles" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150815 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b88d73-0f40-457f-8ece-5cdb2b87a0ec" containerName="collect-profiles" Oct 04 05:00:10 crc kubenswrapper[4575]: E1004 05:00:10.150827 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="854288d3-ecc2-482f-a986-f27cc7672efa" containerName="dnsmasq-dns" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.150834 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="854288d3-ecc2-482f-a986-f27cc7672efa" containerName="dnsmasq-dns" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.151060 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerName="nova-api-api" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.151076 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="854288d3-ecc2-482f-a986-f27cc7672efa" containerName="dnsmasq-dns" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.151092 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="ceilometer-central-agent" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.151102 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b88d73-0f40-457f-8ece-5cdb2b87a0ec" containerName="collect-profiles" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.151113 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="proxy-httpd" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.151131 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcf3c083-541f-4717-9fb1-620a4efb749f" containerName="nova-api-log" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.151140 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="ceilometer-notification-agent" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.151154 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed61f36b-b089-44aa-a413-616112487044" containerName="sg-core" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.151167 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d863756-7f8d-4e1c-ad16-b44be09678c1" containerName="nova-manage" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.152406 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.167109 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.169426 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.169436 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.169648 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.209556 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.212320 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.216792 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.217093 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.234692 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.257865 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.264031 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-public-tls-certs\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.264106 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92kjq\" (UniqueName: \"kubernetes.io/projected/8edac101-a91c-4247-a428-53243b6e5c8e-kube-api-access-92kjq\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.264141 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.264161 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.264205 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-config-data\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.264231 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8edac101-a91c-4247-a428-53243b6e5c8e-logs\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365682 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365746 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-config-data\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365781 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8edac101-a91c-4247-a428-53243b6e5c8e-logs\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365801 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vg4w\" (UniqueName: \"kubernetes.io/projected/36877e92-8c86-4ccd-848c-494f46963048-kube-api-access-6vg4w\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365841 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-config-data\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365865 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36877e92-8c86-4ccd-848c-494f46963048-log-httpd\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365901 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-public-tls-certs\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365931 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-scripts\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365952 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.365986 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36877e92-8c86-4ccd-848c-494f46963048-run-httpd\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.366027 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92kjq\" (UniqueName: \"kubernetes.io/projected/8edac101-a91c-4247-a428-53243b6e5c8e-kube-api-access-92kjq\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.366055 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.366091 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.366120 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.366413 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8edac101-a91c-4247-a428-53243b6e5c8e-logs\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.375394 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.376149 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-config-data\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.376515 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-public-tls-certs\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.377232 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8edac101-a91c-4247-a428-53243b6e5c8e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.404286 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92kjq\" (UniqueName: \"kubernetes.io/projected/8edac101-a91c-4247-a428-53243b6e5c8e-kube-api-access-92kjq\") pod \"nova-api-0\" (UID: \"8edac101-a91c-4247-a428-53243b6e5c8e\") " pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.468851 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.468948 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.469011 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vg4w\" (UniqueName: \"kubernetes.io/projected/36877e92-8c86-4ccd-848c-494f46963048-kube-api-access-6vg4w\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.469064 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-config-data\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.469090 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36877e92-8c86-4ccd-848c-494f46963048-log-httpd\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.469145 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-scripts\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.469166 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.469214 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36877e92-8c86-4ccd-848c-494f46963048-run-httpd\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.469709 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36877e92-8c86-4ccd-848c-494f46963048-run-httpd\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.469951 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36877e92-8c86-4ccd-848c-494f46963048-log-httpd\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.476466 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-config-data\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.476828 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.480461 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.496072 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-scripts\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.497995 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.498970 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36877e92-8c86-4ccd-848c-494f46963048-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.522111 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vg4w\" (UniqueName: \"kubernetes.io/projected/36877e92-8c86-4ccd-848c-494f46963048-kube-api-access-6vg4w\") pod \"ceilometer-0\" (UID: \"36877e92-8c86-4ccd-848c-494f46963048\") " pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.533337 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.746054 4575 generic.go:334] "Generic (PLEG): container finished" podID="a10047f1-cb48-4f95-a18b-2dd340669569" containerID="70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7" exitCode=143 Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.746411 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a10047f1-cb48-4f95-a18b-2dd340669569","Type":"ContainerDied","Data":"70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7"} Oct 04 05:00:10 crc kubenswrapper[4575]: I1004 05:00:10.755864 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dtc62" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="registry-server" containerID="cri-o://3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459" gracePeriod=2 Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.084669 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.216026 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 05:00:11 crc kubenswrapper[4575]: W1004 05:00:11.237110 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36877e92_8c86_4ccd_848c_494f46963048.slice/crio-49fb5f8a32ce4f0802497ce7acedfc41dbb3d6b677b1214f8cad09eefd92c6ca WatchSource:0}: Error finding container 49fb5f8a32ce4f0802497ce7acedfc41dbb3d6b677b1214f8cad09eefd92c6ca: Status 404 returned error can't find the container with id 49fb5f8a32ce4f0802497ce7acedfc41dbb3d6b677b1214f8cad09eefd92c6ca Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.326152 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcf3c083-541f-4717-9fb1-620a4efb749f" path="/var/lib/kubelet/pods/bcf3c083-541f-4717-9fb1-620a4efb749f/volumes" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.347154 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed61f36b-b089-44aa-a413-616112487044" path="/var/lib/kubelet/pods/ed61f36b-b089-44aa-a413-616112487044/volumes" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.350025 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.480649 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.504088 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-catalog-content\") pod \"c231de36-c828-4591-affb-cf345a773289\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.504284 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj2b2\" (UniqueName: \"kubernetes.io/projected/c231de36-c828-4591-affb-cf345a773289-kube-api-access-nj2b2\") pod \"c231de36-c828-4591-affb-cf345a773289\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.504310 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-utilities\") pod \"c231de36-c828-4591-affb-cf345a773289\" (UID: \"c231de36-c828-4591-affb-cf345a773289\") " Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.511152 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-utilities" (OuterVolumeSpecName: "utilities") pod "c231de36-c828-4591-affb-cf345a773289" (UID: "c231de36-c828-4591-affb-cf345a773289"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.524332 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c231de36-c828-4591-affb-cf345a773289-kube-api-access-nj2b2" (OuterVolumeSpecName: "kube-api-access-nj2b2") pod "c231de36-c828-4591-affb-cf345a773289" (UID: "c231de36-c828-4591-affb-cf345a773289"). InnerVolumeSpecName "kube-api-access-nj2b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.604745 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c231de36-c828-4591-affb-cf345a773289" (UID: "c231de36-c828-4591-affb-cf345a773289"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.606954 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-config-data\") pod \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.607178 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-combined-ca-bundle\") pod \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.607282 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65xnk\" (UniqueName: \"kubernetes.io/projected/15a3821f-f7ca-45b4-b667-d5985bc25f3a-kube-api-access-65xnk\") pod \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\" (UID: \"15a3821f-f7ca-45b4-b667-d5985bc25f3a\") " Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.607990 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj2b2\" (UniqueName: \"kubernetes.io/projected/c231de36-c828-4591-affb-cf345a773289-kube-api-access-nj2b2\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.608009 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.608021 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c231de36-c828-4591-affb-cf345a773289-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.612200 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a3821f-f7ca-45b4-b667-d5985bc25f3a-kube-api-access-65xnk" (OuterVolumeSpecName: "kube-api-access-65xnk") pod "15a3821f-f7ca-45b4-b667-d5985bc25f3a" (UID: "15a3821f-f7ca-45b4-b667-d5985bc25f3a"). InnerVolumeSpecName "kube-api-access-65xnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.647007 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-config-data" (OuterVolumeSpecName: "config-data") pod "15a3821f-f7ca-45b4-b667-d5985bc25f3a" (UID: "15a3821f-f7ca-45b4-b667-d5985bc25f3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.663479 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15a3821f-f7ca-45b4-b667-d5985bc25f3a" (UID: "15a3821f-f7ca-45b4-b667-d5985bc25f3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.709757 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.710011 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a3821f-f7ca-45b4-b667-d5985bc25f3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.710102 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65xnk\" (UniqueName: \"kubernetes.io/projected/15a3821f-f7ca-45b4-b667-d5985bc25f3a-kube-api-access-65xnk\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.771308 4575 generic.go:334] "Generic (PLEG): container finished" podID="15a3821f-f7ca-45b4-b667-d5985bc25f3a" containerID="c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c" exitCode=0 Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.771366 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15a3821f-f7ca-45b4-b667-d5985bc25f3a","Type":"ContainerDied","Data":"c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c"} Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.771410 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.771649 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15a3821f-f7ca-45b4-b667-d5985bc25f3a","Type":"ContainerDied","Data":"bd87db666666b645584df8cbb413b693281a018585f08eaa9bbe2522ee052e07"} Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.771686 4575 scope.go:117] "RemoveContainer" containerID="c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.775762 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8edac101-a91c-4247-a428-53243b6e5c8e","Type":"ContainerStarted","Data":"a0a85ae7de7397a0aa12ebb680e6adc94c80a65e97986c5ae051694524b04cf5"} Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.775795 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8edac101-a91c-4247-a428-53243b6e5c8e","Type":"ContainerStarted","Data":"cf1f0a962454d8b1de01f391d1067c02472ecd107c9883a61641aa906a2de1cc"} Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.775805 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8edac101-a91c-4247-a428-53243b6e5c8e","Type":"ContainerStarted","Data":"7e0d7b813cc5342c91bf5992db4e4f8df7609f49a013f4866006dd7e1bb07faa"} Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.779734 4575 generic.go:334] "Generic (PLEG): container finished" podID="c231de36-c828-4591-affb-cf345a773289" containerID="3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459" exitCode=0 Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.779861 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtc62" event={"ID":"c231de36-c828-4591-affb-cf345a773289","Type":"ContainerDied","Data":"3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459"} Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.779898 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtc62" event={"ID":"c231de36-c828-4591-affb-cf345a773289","Type":"ContainerDied","Data":"edd570faf2fef63d70c937024afd8e4766fd090ad402fd7899d06fa54120270d"} Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.779977 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtc62" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.786651 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36877e92-8c86-4ccd-848c-494f46963048","Type":"ContainerStarted","Data":"49fb5f8a32ce4f0802497ce7acedfc41dbb3d6b677b1214f8cad09eefd92c6ca"} Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.797425 4575 scope.go:117] "RemoveContainer" containerID="c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c" Oct 04 05:00:11 crc kubenswrapper[4575]: E1004 05:00:11.801356 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c\": container with ID starting with c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c not found: ID does not exist" containerID="c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.804978 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c"} err="failed to get container status \"c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c\": rpc error: code = NotFound desc = could not find container \"c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c\": container with ID starting with c0205512cfa084c6ba1a4df8e6d93f0b3bba73d7735891d40aa79b09d033aa9c not found: ID does not exist" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.805013 4575 scope.go:117] "RemoveContainer" containerID="3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.818948 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.831871 4575 scope.go:117] "RemoveContainer" containerID="72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.846732 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.860718 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dtc62"] Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.870034 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dtc62"] Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.875857 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:00:11 crc kubenswrapper[4575]: E1004 05:00:11.876319 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="registry-server" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.876337 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="registry-server" Oct 04 05:00:11 crc kubenswrapper[4575]: E1004 05:00:11.876381 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="extract-utilities" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.876388 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="extract-utilities" Oct 04 05:00:11 crc kubenswrapper[4575]: E1004 05:00:11.876400 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a3821f-f7ca-45b4-b667-d5985bc25f3a" containerName="nova-scheduler-scheduler" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.876406 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a3821f-f7ca-45b4-b667-d5985bc25f3a" containerName="nova-scheduler-scheduler" Oct 04 05:00:11 crc kubenswrapper[4575]: E1004 05:00:11.876423 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="extract-content" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.876430 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="extract-content" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.876623 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="15a3821f-f7ca-45b4-b667-d5985bc25f3a" containerName="nova-scheduler-scheduler" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.876641 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c231de36-c828-4591-affb-cf345a773289" containerName="registry-server" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.877285 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.883808 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.885850 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:00:11 crc kubenswrapper[4575]: I1004 05:00:11.954820 4575 scope.go:117] "RemoveContainer" containerID="406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.016853 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2f74\" (UniqueName: \"kubernetes.io/projected/fe22583c-fe25-4a83-a80b-23670a8cf79a-kube-api-access-b2f74\") pod \"nova-scheduler-0\" (UID: \"fe22583c-fe25-4a83-a80b-23670a8cf79a\") " pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.016901 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe22583c-fe25-4a83-a80b-23670a8cf79a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fe22583c-fe25-4a83-a80b-23670a8cf79a\") " pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.016935 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe22583c-fe25-4a83-a80b-23670a8cf79a-config-data\") pod \"nova-scheduler-0\" (UID: \"fe22583c-fe25-4a83-a80b-23670a8cf79a\") " pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.088127 4575 scope.go:117] "RemoveContainer" containerID="3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459" Oct 04 05:00:12 crc kubenswrapper[4575]: E1004 05:00:12.089945 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459\": container with ID starting with 3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459 not found: ID does not exist" containerID="3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.089990 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459"} err="failed to get container status \"3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459\": rpc error: code = NotFound desc = could not find container \"3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459\": container with ID starting with 3cd5464de4e8949a7306ca678b0b2678f8b654813ff442416444e19768074459 not found: ID does not exist" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.090016 4575 scope.go:117] "RemoveContainer" containerID="72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a" Oct 04 05:00:12 crc kubenswrapper[4575]: E1004 05:00:12.090499 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a\": container with ID starting with 72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a not found: ID does not exist" containerID="72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.090534 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a"} err="failed to get container status \"72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a\": rpc error: code = NotFound desc = could not find container \"72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a\": container with ID starting with 72d3f769378bd2c7d7629f183e1a2e4fe21d6449b1b7c0fe53f19d087c9afc4a not found: ID does not exist" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.090553 4575 scope.go:117] "RemoveContainer" containerID="406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea" Oct 04 05:00:12 crc kubenswrapper[4575]: E1004 05:00:12.090897 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea\": container with ID starting with 406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea not found: ID does not exist" containerID="406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.090938 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea"} err="failed to get container status \"406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea\": rpc error: code = NotFound desc = could not find container \"406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea\": container with ID starting with 406b230452d68f004fbeb493108e1c7615101b27e6e1f02cf2f937f033aca0ea not found: ID does not exist" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.119485 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2f74\" (UniqueName: \"kubernetes.io/projected/fe22583c-fe25-4a83-a80b-23670a8cf79a-kube-api-access-b2f74\") pod \"nova-scheduler-0\" (UID: \"fe22583c-fe25-4a83-a80b-23670a8cf79a\") " pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.119543 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe22583c-fe25-4a83-a80b-23670a8cf79a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fe22583c-fe25-4a83-a80b-23670a8cf79a\") " pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.119629 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe22583c-fe25-4a83-a80b-23670a8cf79a-config-data\") pod \"nova-scheduler-0\" (UID: \"fe22583c-fe25-4a83-a80b-23670a8cf79a\") " pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.123295 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe22583c-fe25-4a83-a80b-23670a8cf79a-config-data\") pod \"nova-scheduler-0\" (UID: \"fe22583c-fe25-4a83-a80b-23670a8cf79a\") " pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.124621 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe22583c-fe25-4a83-a80b-23670a8cf79a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fe22583c-fe25-4a83-a80b-23670a8cf79a\") " pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.141338 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2f74\" (UniqueName: \"kubernetes.io/projected/fe22583c-fe25-4a83-a80b-23670a8cf79a-kube-api-access-b2f74\") pod \"nova-scheduler-0\" (UID: \"fe22583c-fe25-4a83-a80b-23670a8cf79a\") " pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.201182 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.641678 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:00:12 crc kubenswrapper[4575]: W1004 05:00:12.643098 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe22583c_fe25_4a83_a80b_23670a8cf79a.slice/crio-ff6bc8ec90dd323d3bb077d8ee9a9299c0e102f2ec1c13653a100149bbedfe19 WatchSource:0}: Error finding container ff6bc8ec90dd323d3bb077d8ee9a9299c0e102f2ec1c13653a100149bbedfe19: Status 404 returned error can't find the container with id ff6bc8ec90dd323d3bb077d8ee9a9299c0e102f2ec1c13653a100149bbedfe19 Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.798361 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36877e92-8c86-4ccd-848c-494f46963048","Type":"ContainerStarted","Data":"e5456aef0033bfb268b0330a055a6a14182f6e5bea55c030e35bed07017de30e"} Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.801519 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fe22583c-fe25-4a83-a80b-23670a8cf79a","Type":"ContainerStarted","Data":"ff6bc8ec90dd323d3bb077d8ee9a9299c0e102f2ec1c13653a100149bbedfe19"} Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.829402 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.829378941 podStartE2EDuration="2.829378941s" podCreationTimestamp="2025-10-04 05:00:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:00:12.818149665 +0000 UTC m=+1564.146708499" watchObservedRunningTime="2025-10-04 05:00:12.829378941 +0000 UTC m=+1564.157937755" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.851507 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:42992->10.217.0.198:8775: read: connection reset by peer" Oct 04 05:00:12 crc kubenswrapper[4575]: I1004 05:00:12.851506 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:42994->10.217.0.198:8775: read: connection reset by peer" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.353305 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15a3821f-f7ca-45b4-b667-d5985bc25f3a" path="/var/lib/kubelet/pods/15a3821f-f7ca-45b4-b667-d5985bc25f3a/volumes" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.354691 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c231de36-c828-4591-affb-cf345a773289" path="/var/lib/kubelet/pods/c231de36-c828-4591-affb-cf345a773289/volumes" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.363750 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.459880 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrkk6\" (UniqueName: \"kubernetes.io/projected/a10047f1-cb48-4f95-a18b-2dd340669569-kube-api-access-xrkk6\") pod \"a10047f1-cb48-4f95-a18b-2dd340669569\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.459973 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-config-data\") pod \"a10047f1-cb48-4f95-a18b-2dd340669569\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.460353 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-nova-metadata-tls-certs\") pod \"a10047f1-cb48-4f95-a18b-2dd340669569\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.460474 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10047f1-cb48-4f95-a18b-2dd340669569-logs\") pod \"a10047f1-cb48-4f95-a18b-2dd340669569\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.460630 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-combined-ca-bundle\") pod \"a10047f1-cb48-4f95-a18b-2dd340669569\" (UID: \"a10047f1-cb48-4f95-a18b-2dd340669569\") " Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.463626 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a10047f1-cb48-4f95-a18b-2dd340669569-logs" (OuterVolumeSpecName: "logs") pod "a10047f1-cb48-4f95-a18b-2dd340669569" (UID: "a10047f1-cb48-4f95-a18b-2dd340669569"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.488981 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a10047f1-cb48-4f95-a18b-2dd340669569-kube-api-access-xrkk6" (OuterVolumeSpecName: "kube-api-access-xrkk6") pod "a10047f1-cb48-4f95-a18b-2dd340669569" (UID: "a10047f1-cb48-4f95-a18b-2dd340669569"). InnerVolumeSpecName "kube-api-access-xrkk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.535052 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-config-data" (OuterVolumeSpecName: "config-data") pod "a10047f1-cb48-4f95-a18b-2dd340669569" (UID: "a10047f1-cb48-4f95-a18b-2dd340669569"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.543580 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a10047f1-cb48-4f95-a18b-2dd340669569" (UID: "a10047f1-cb48-4f95-a18b-2dd340669569"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.564410 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10047f1-cb48-4f95-a18b-2dd340669569-logs\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.564453 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.564482 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrkk6\" (UniqueName: \"kubernetes.io/projected/a10047f1-cb48-4f95-a18b-2dd340669569-kube-api-access-xrkk6\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.564494 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.587904 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a10047f1-cb48-4f95-a18b-2dd340669569" (UID: "a10047f1-cb48-4f95-a18b-2dd340669569"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.666289 4575 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a10047f1-cb48-4f95-a18b-2dd340669569-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.812980 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fe22583c-fe25-4a83-a80b-23670a8cf79a","Type":"ContainerStarted","Data":"fff4d4fed20ae6b5296f1da504eb04c45bff270512444efa2c11a86771920989"} Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.815334 4575 generic.go:334] "Generic (PLEG): container finished" podID="a10047f1-cb48-4f95-a18b-2dd340669569" containerID="7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3" exitCode=0 Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.815441 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a10047f1-cb48-4f95-a18b-2dd340669569","Type":"ContainerDied","Data":"7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3"} Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.815483 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a10047f1-cb48-4f95-a18b-2dd340669569","Type":"ContainerDied","Data":"cf628e9ae6e08ed7de73939d19e485ebb3861e9ac56f0408de20b38c93799baa"} Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.815506 4575 scope.go:117] "RemoveContainer" containerID="7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.815506 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.817992 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36877e92-8c86-4ccd-848c-494f46963048","Type":"ContainerStarted","Data":"7488de6e62c4f48279e8e884e9f77ffcd7d0cdf62ec23b2b539ace9bd597e05a"} Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.848169 4575 scope.go:117] "RemoveContainer" containerID="70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.850823 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.850790198 podStartE2EDuration="2.850790198s" podCreationTimestamp="2025-10-04 05:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:00:13.833622279 +0000 UTC m=+1565.162181103" watchObservedRunningTime="2025-10-04 05:00:13.850790198 +0000 UTC m=+1565.179349012" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.873746 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.875534 4575 scope.go:117] "RemoveContainer" containerID="7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3" Oct 04 05:00:13 crc kubenswrapper[4575]: E1004 05:00:13.877783 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3\": container with ID starting with 7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3 not found: ID does not exist" containerID="7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.877827 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3"} err="failed to get container status \"7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3\": rpc error: code = NotFound desc = could not find container \"7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3\": container with ID starting with 7e54de879bab52248e9b677911d8ead938f3ff935c37e48b6050398c48cd83f3 not found: ID does not exist" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.877852 4575 scope.go:117] "RemoveContainer" containerID="70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7" Oct 04 05:00:13 crc kubenswrapper[4575]: E1004 05:00:13.880247 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7\": container with ID starting with 70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7 not found: ID does not exist" containerID="70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.880286 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7"} err="failed to get container status \"70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7\": rpc error: code = NotFound desc = could not find container \"70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7\": container with ID starting with 70551ba66c3f423ca4a9f0a743995493b3f3a6c311f46e539d3efa98ce9641f7 not found: ID does not exist" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.889304 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.898570 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:00:13 crc kubenswrapper[4575]: E1004 05:00:13.899113 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-log" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.899140 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-log" Oct 04 05:00:13 crc kubenswrapper[4575]: E1004 05:00:13.899158 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-metadata" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.899171 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-metadata" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.899410 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-metadata" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.899432 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" containerName="nova-metadata-log" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.900662 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.902823 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.903106 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.910571 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.978487 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7664c119-13b3-49a5-bd83-ba72b8d54c07-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.979057 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prgtb\" (UniqueName: \"kubernetes.io/projected/7664c119-13b3-49a5-bd83-ba72b8d54c07-kube-api-access-prgtb\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.979126 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7664c119-13b3-49a5-bd83-ba72b8d54c07-config-data\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.979174 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7664c119-13b3-49a5-bd83-ba72b8d54c07-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:13 crc kubenswrapper[4575]: I1004 05:00:13.979307 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7664c119-13b3-49a5-bd83-ba72b8d54c07-logs\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.081297 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7664c119-13b3-49a5-bd83-ba72b8d54c07-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.081381 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7664c119-13b3-49a5-bd83-ba72b8d54c07-logs\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.081438 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7664c119-13b3-49a5-bd83-ba72b8d54c07-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.081492 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prgtb\" (UniqueName: \"kubernetes.io/projected/7664c119-13b3-49a5-bd83-ba72b8d54c07-kube-api-access-prgtb\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.081537 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7664c119-13b3-49a5-bd83-ba72b8d54c07-config-data\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.081811 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7664c119-13b3-49a5-bd83-ba72b8d54c07-logs\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.101329 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7664c119-13b3-49a5-bd83-ba72b8d54c07-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.104007 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7664c119-13b3-49a5-bd83-ba72b8d54c07-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.107986 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7664c119-13b3-49a5-bd83-ba72b8d54c07-config-data\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.110135 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prgtb\" (UniqueName: \"kubernetes.io/projected/7664c119-13b3-49a5-bd83-ba72b8d54c07-kube-api-access-prgtb\") pod \"nova-metadata-0\" (UID: \"7664c119-13b3-49a5-bd83-ba72b8d54c07\") " pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.220167 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.601254 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.833175 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7664c119-13b3-49a5-bd83-ba72b8d54c07","Type":"ContainerStarted","Data":"83c5200d5f2e92fc89ff653ea30fec7643d0c7a01fee3c0a8bc354a3f3e74bda"} Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.833455 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7664c119-13b3-49a5-bd83-ba72b8d54c07","Type":"ContainerStarted","Data":"fba79784cf77e08b7377e9dd753c5cbe41cb35944c117bc694cf46897b5fb03d"} Oct 04 05:00:14 crc kubenswrapper[4575]: I1004 05:00:14.841966 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36877e92-8c86-4ccd-848c-494f46963048","Type":"ContainerStarted","Data":"784e3abf2c4c9d78cdad1244e0031bbd2e8edca7f7e8fd67f1c6abbb92edcff1"} Oct 04 05:00:15 crc kubenswrapper[4575]: I1004 05:00:15.321818 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a10047f1-cb48-4f95-a18b-2dd340669569" path="/var/lib/kubelet/pods/a10047f1-cb48-4f95-a18b-2dd340669569/volumes" Oct 04 05:00:15 crc kubenswrapper[4575]: I1004 05:00:15.854449 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7664c119-13b3-49a5-bd83-ba72b8d54c07","Type":"ContainerStarted","Data":"ba86146fe034df6dc42fbbd787a083e6a8f5a6da74bf8edaa47822f653b7ff17"} Oct 04 05:00:15 crc kubenswrapper[4575]: I1004 05:00:15.876941 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.876923279 podStartE2EDuration="2.876923279s" podCreationTimestamp="2025-10-04 05:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:00:15.870164482 +0000 UTC m=+1567.198723296" watchObservedRunningTime="2025-10-04 05:00:15.876923279 +0000 UTC m=+1567.205482113" Oct 04 05:00:16 crc kubenswrapper[4575]: I1004 05:00:16.885600 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36877e92-8c86-4ccd-848c-494f46963048","Type":"ContainerStarted","Data":"5d3c571728772d23ebbda2f2eb68746f649a3d1ed781e47289d6f9e85814c056"} Oct 04 05:00:16 crc kubenswrapper[4575]: I1004 05:00:16.886478 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 05:00:16 crc kubenswrapper[4575]: I1004 05:00:16.920702 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.206483288 podStartE2EDuration="6.920682415s" podCreationTimestamp="2025-10-04 05:00:10 +0000 UTC" firstStartedPulling="2025-10-04 05:00:11.272466029 +0000 UTC m=+1562.601024843" lastFinishedPulling="2025-10-04 05:00:15.986665156 +0000 UTC m=+1567.315223970" observedRunningTime="2025-10-04 05:00:16.913552347 +0000 UTC m=+1568.242111171" watchObservedRunningTime="2025-10-04 05:00:16.920682415 +0000 UTC m=+1568.249241229" Oct 04 05:00:17 crc kubenswrapper[4575]: I1004 05:00:17.202209 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 05:00:19 crc kubenswrapper[4575]: I1004 05:00:19.220937 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 05:00:19 crc kubenswrapper[4575]: I1004 05:00:19.221289 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 05:00:20 crc kubenswrapper[4575]: I1004 05:00:20.481406 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 05:00:20 crc kubenswrapper[4575]: I1004 05:00:20.481859 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 05:00:21 crc kubenswrapper[4575]: I1004 05:00:21.494878 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8edac101-a91c-4247-a428-53243b6e5c8e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 05:00:21 crc kubenswrapper[4575]: I1004 05:00:21.494916 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8edac101-a91c-4247-a428-53243b6e5c8e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 05:00:22 crc kubenswrapper[4575]: I1004 05:00:22.201974 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 05:00:22 crc kubenswrapper[4575]: I1004 05:00:22.232737 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 05:00:22 crc kubenswrapper[4575]: I1004 05:00:22.965210 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 05:00:24 crc kubenswrapper[4575]: I1004 05:00:24.221165 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 05:00:24 crc kubenswrapper[4575]: I1004 05:00:24.221215 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 05:00:25 crc kubenswrapper[4575]: I1004 05:00:25.233937 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7664c119-13b3-49a5-bd83-ba72b8d54c07" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 05:00:25 crc kubenswrapper[4575]: I1004 05:00:25.233954 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7664c119-13b3-49a5-bd83-ba72b8d54c07" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 05:00:30 crc kubenswrapper[4575]: I1004 05:00:30.488538 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 05:00:30 crc kubenswrapper[4575]: I1004 05:00:30.489361 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 05:00:30 crc kubenswrapper[4575]: I1004 05:00:30.490438 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 05:00:30 crc kubenswrapper[4575]: I1004 05:00:30.495997 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 05:00:31 crc kubenswrapper[4575]: I1004 05:00:31.018156 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 05:00:31 crc kubenswrapper[4575]: I1004 05:00:31.026218 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 05:00:34 crc kubenswrapper[4575]: I1004 05:00:34.226291 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 05:00:34 crc kubenswrapper[4575]: I1004 05:00:34.230873 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 05:00:34 crc kubenswrapper[4575]: I1004 05:00:34.234193 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 05:00:35 crc kubenswrapper[4575]: I1004 05:00:35.058053 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 05:00:36 crc kubenswrapper[4575]: I1004 05:00:36.063974 4575 generic.go:334] "Generic (PLEG): container finished" podID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerID="f8d4443c36b0c2245629253c64c8178d2ca6381e985a3cf457f5457982a4dcd8" exitCode=137 Oct 04 05:00:36 crc kubenswrapper[4575]: I1004 05:00:36.064031 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerDied","Data":"f8d4443c36b0c2245629253c64c8178d2ca6381e985a3cf457f5457982a4dcd8"} Oct 04 05:00:36 crc kubenswrapper[4575]: I1004 05:00:36.064849 4575 scope.go:117] "RemoveContainer" containerID="ec938e12e0dcd7b18a29f344980b6af8ff56070e067b69f96ecbd149ba683cb0" Oct 04 05:00:36 crc kubenswrapper[4575]: I1004 05:00:36.067539 4575 generic.go:334] "Generic (PLEG): container finished" podID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerID="1286c3d99af99d1bd1f66544432932ee55a667d406d520ce4179dfad08a0c438" exitCode=137 Oct 04 05:00:36 crc kubenswrapper[4575]: I1004 05:00:36.067655 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerDied","Data":"1286c3d99af99d1bd1f66544432932ee55a667d406d520ce4179dfad08a0c438"} Oct 04 05:00:36 crc kubenswrapper[4575]: I1004 05:00:36.460679 4575 scope.go:117] "RemoveContainer" containerID="376a8e8470b9959c733967cacb7a6d6b2b157401d6330173cf7a8725128fbf8f" Oct 04 05:00:37 crc kubenswrapper[4575]: I1004 05:00:37.079357 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-644bf5cdd4-pwdw9" event={"ID":"6f0573f0-1418-43eb-be82-166fb99ed23d","Type":"ContainerStarted","Data":"79b7ecfa0da7cccae19c2ef6d39ffb1f37267a4e8d6ba510c85809412f567078"} Oct 04 05:00:37 crc kubenswrapper[4575]: I1004 05:00:37.082290 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerStarted","Data":"ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b"} Oct 04 05:00:38 crc kubenswrapper[4575]: I1004 05:00:38.446241 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:00:38 crc kubenswrapper[4575]: I1004 05:00:38.446598 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:00:40 crc kubenswrapper[4575]: I1004 05:00:40.548035 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 05:00:45 crc kubenswrapper[4575]: I1004 05:00:45.557386 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 05:00:45 crc kubenswrapper[4575]: I1004 05:00:45.557920 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 05:00:45 crc kubenswrapper[4575]: I1004 05:00:45.713733 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 05:00:45 crc kubenswrapper[4575]: I1004 05:00:45.713818 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 05:00:55 crc kubenswrapper[4575]: I1004 05:00:55.559277 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 05:00:55 crc kubenswrapper[4575]: I1004 05:00:55.714995 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-644bf5cdd4-pwdw9" podUID="6f0573f0-1418-43eb-be82-166fb99ed23d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.165409 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325901-nx5vx"] Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.167097 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.186696 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325901-nx5vx"] Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.209443 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-fernet-keys\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.209739 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-combined-ca-bundle\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.209937 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cthcf\" (UniqueName: \"kubernetes.io/projected/4c0652b0-f1a2-4439-87b1-2a083fdbb535-kube-api-access-cthcf\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.210063 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-config-data\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.311448 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cthcf\" (UniqueName: \"kubernetes.io/projected/4c0652b0-f1a2-4439-87b1-2a083fdbb535-kube-api-access-cthcf\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.311557 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-config-data\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.311886 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-fernet-keys\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.311921 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-combined-ca-bundle\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.320541 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-combined-ca-bundle\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.328628 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-fernet-keys\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.331853 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-config-data\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.334974 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cthcf\" (UniqueName: \"kubernetes.io/projected/4c0652b0-f1a2-4439-87b1-2a083fdbb535-kube-api-access-cthcf\") pod \"keystone-cron-29325901-nx5vx\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:00 crc kubenswrapper[4575]: I1004 05:01:00.490029 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:01 crc kubenswrapper[4575]: I1004 05:01:01.053492 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325901-nx5vx"] Oct 04 05:01:01 crc kubenswrapper[4575]: I1004 05:01:01.307958 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-nx5vx" event={"ID":"4c0652b0-f1a2-4439-87b1-2a083fdbb535","Type":"ContainerStarted","Data":"1ddbcb4e3380c7ae4534a896d1c870ded54c725cba08792a5acfd867c0ec91c5"} Oct 04 05:01:01 crc kubenswrapper[4575]: I1004 05:01:01.309448 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-nx5vx" event={"ID":"4c0652b0-f1a2-4439-87b1-2a083fdbb535","Type":"ContainerStarted","Data":"fa71242dac69fbc9383a913c89eab4c52563e955f82d31e147f464ba6bce1975"} Oct 04 05:01:01 crc kubenswrapper[4575]: I1004 05:01:01.337553 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325901-nx5vx" podStartSLOduration=1.337532159 podStartE2EDuration="1.337532159s" podCreationTimestamp="2025-10-04 05:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:01:01.337072595 +0000 UTC m=+1612.665631429" watchObservedRunningTime="2025-10-04 05:01:01.337532159 +0000 UTC m=+1612.666090973" Oct 04 05:01:04 crc kubenswrapper[4575]: I1004 05:01:04.340054 4575 generic.go:334] "Generic (PLEG): container finished" podID="4c0652b0-f1a2-4439-87b1-2a083fdbb535" containerID="1ddbcb4e3380c7ae4534a896d1c870ded54c725cba08792a5acfd867c0ec91c5" exitCode=0 Oct 04 05:01:04 crc kubenswrapper[4575]: I1004 05:01:04.340240 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-nx5vx" event={"ID":"4c0652b0-f1a2-4439-87b1-2a083fdbb535","Type":"ContainerDied","Data":"1ddbcb4e3380c7ae4534a896d1c870ded54c725cba08792a5acfd867c0ec91c5"} Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.706537 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.721981 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cthcf\" (UniqueName: \"kubernetes.io/projected/4c0652b0-f1a2-4439-87b1-2a083fdbb535-kube-api-access-cthcf\") pod \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.722064 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-config-data\") pod \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.722101 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-combined-ca-bundle\") pod \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.722178 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-fernet-keys\") pod \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\" (UID: \"4c0652b0-f1a2-4439-87b1-2a083fdbb535\") " Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.730248 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c0652b0-f1a2-4439-87b1-2a083fdbb535-kube-api-access-cthcf" (OuterVolumeSpecName: "kube-api-access-cthcf") pod "4c0652b0-f1a2-4439-87b1-2a083fdbb535" (UID: "4c0652b0-f1a2-4439-87b1-2a083fdbb535"). InnerVolumeSpecName "kube-api-access-cthcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.742682 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4c0652b0-f1a2-4439-87b1-2a083fdbb535" (UID: "4c0652b0-f1a2-4439-87b1-2a083fdbb535"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.780666 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c0652b0-f1a2-4439-87b1-2a083fdbb535" (UID: "4c0652b0-f1a2-4439-87b1-2a083fdbb535"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.812171 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-config-data" (OuterVolumeSpecName: "config-data") pod "4c0652b0-f1a2-4439-87b1-2a083fdbb535" (UID: "4c0652b0-f1a2-4439-87b1-2a083fdbb535"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.824197 4575 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.824228 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cthcf\" (UniqueName: \"kubernetes.io/projected/4c0652b0-f1a2-4439-87b1-2a083fdbb535-kube-api-access-cthcf\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.824242 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:05 crc kubenswrapper[4575]: I1004 05:01:05.824252 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c0652b0-f1a2-4439-87b1-2a083fdbb535-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:06 crc kubenswrapper[4575]: I1004 05:01:06.358466 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-nx5vx" event={"ID":"4c0652b0-f1a2-4439-87b1-2a083fdbb535","Type":"ContainerDied","Data":"fa71242dac69fbc9383a913c89eab4c52563e955f82d31e147f464ba6bce1975"} Oct 04 05:01:06 crc kubenswrapper[4575]: I1004 05:01:06.358814 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa71242dac69fbc9383a913c89eab4c52563e955f82d31e147f464ba6bce1975" Oct 04 05:01:06 crc kubenswrapper[4575]: I1004 05:01:06.358499 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-nx5vx" Oct 04 05:01:08 crc kubenswrapper[4575]: I1004 05:01:08.027007 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 05:01:08 crc kubenswrapper[4575]: I1004 05:01:08.051475 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 05:01:08 crc kubenswrapper[4575]: I1004 05:01:08.446281 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:01:08 crc kubenswrapper[4575]: I1004 05:01:08.446339 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:01:08 crc kubenswrapper[4575]: I1004 05:01:08.446380 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:01:08 crc kubenswrapper[4575]: I1004 05:01:08.447129 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:01:08 crc kubenswrapper[4575]: I1004 05:01:08.447203 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" gracePeriod=600 Oct 04 05:01:08 crc kubenswrapper[4575]: E1004 05:01:08.578554 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:01:09 crc kubenswrapper[4575]: I1004 05:01:09.385540 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" exitCode=0 Oct 04 05:01:09 crc kubenswrapper[4575]: I1004 05:01:09.385609 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22"} Oct 04 05:01:09 crc kubenswrapper[4575]: I1004 05:01:09.385960 4575 scope.go:117] "RemoveContainer" containerID="fec7eca9bd28e99b532e638e857f42e7eb7fe8c1e89c0dd52e522352fd86d19d" Oct 04 05:01:09 crc kubenswrapper[4575]: I1004 05:01:09.394051 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:01:09 crc kubenswrapper[4575]: E1004 05:01:09.394471 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:01:09 crc kubenswrapper[4575]: I1004 05:01:09.802274 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 05:01:09 crc kubenswrapper[4575]: I1004 05:01:09.808547 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-644bf5cdd4-pwdw9" Oct 04 05:01:09 crc kubenswrapper[4575]: I1004 05:01:09.908921 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b86b879b4-tlf42"] Oct 04 05:01:10 crc kubenswrapper[4575]: I1004 05:01:10.412187 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon-log" containerID="cri-o://bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149" gracePeriod=30 Oct 04 05:01:10 crc kubenswrapper[4575]: I1004 05:01:10.412269 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" containerID="cri-o://ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b" gracePeriod=30 Oct 04 05:01:14 crc kubenswrapper[4575]: I1004 05:01:14.449421 4575 generic.go:334] "Generic (PLEG): container finished" podID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerID="ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b" exitCode=0 Oct 04 05:01:14 crc kubenswrapper[4575]: I1004 05:01:14.449866 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerDied","Data":"ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b"} Oct 04 05:01:14 crc kubenswrapper[4575]: I1004 05:01:14.449996 4575 scope.go:117] "RemoveContainer" containerID="1286c3d99af99d1bd1f66544432932ee55a667d406d520ce4179dfad08a0c438" Oct 04 05:01:15 crc kubenswrapper[4575]: I1004 05:01:15.558073 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 05:01:18 crc kubenswrapper[4575]: I1004 05:01:18.812105 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 05:01:19 crc kubenswrapper[4575]: I1004 05:01:19.533042 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 05:01:22 crc kubenswrapper[4575]: I1004 05:01:22.310719 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:01:22 crc kubenswrapper[4575]: E1004 05:01:22.311308 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:01:23 crc kubenswrapper[4575]: I1004 05:01:23.618884 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" containerName="rabbitmq" containerID="cri-o://827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717" gracePeriod=604796 Oct 04 05:01:24 crc kubenswrapper[4575]: I1004 05:01:24.047214 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" containerName="rabbitmq" containerID="cri-o://2d02ce72a5f63bcbca2e3120fa4a39e38e4cfa8d01de232b93c3fde2b1e34412" gracePeriod=604796 Oct 04 05:01:24 crc kubenswrapper[4575]: I1004 05:01:24.130274 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Oct 04 05:01:24 crc kubenswrapper[4575]: I1004 05:01:24.490248 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 04 05:01:25 crc kubenswrapper[4575]: I1004 05:01:25.557556 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 05:01:27 crc kubenswrapper[4575]: I1004 05:01:27.821267 4575 scope.go:117] "RemoveContainer" containerID="c1934941b3c46049e7df791739c6b54a1c9904ab83f51bff37992df55a51f8c6" Oct 04 05:01:27 crc kubenswrapper[4575]: I1004 05:01:27.847152 4575 scope.go:117] "RemoveContainer" containerID="797b6d89affe9a8940efd0f673e8e991f4a5490d6f2839f11417625c7856ecb7" Oct 04 05:01:27 crc kubenswrapper[4575]: I1004 05:01:27.900714 4575 scope.go:117] "RemoveContainer" containerID="f1617c5123c746de83fa90f0e4e7e10b3e84efd6bd916380ea9cd2369ae67832" Oct 04 05:01:27 crc kubenswrapper[4575]: I1004 05:01:27.940699 4575 scope.go:117] "RemoveContainer" containerID="da60a53726dd43ff90d13c3b3f93974566e2f3b78dfb1904ec524d845f90485e" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.372345 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.502441 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-config-data\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.502510 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-plugins\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.502623 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-server-conf\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.502695 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c16a4e4c-f361-42ef-9af1-e7601b14ca69-pod-info\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.502737 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-plugins-conf\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.502802 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28d8b\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-kube-api-access-28d8b\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.502854 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-erlang-cookie\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.502884 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.502949 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-confd\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.503010 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-tls\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.503040 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c16a4e4c-f361-42ef-9af1-e7601b14ca69-erlang-cookie-secret\") pod \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\" (UID: \"c16a4e4c-f361-42ef-9af1-e7601b14ca69\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.505612 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.511251 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16a4e4c-f361-42ef-9af1-e7601b14ca69-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.520085 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.544513 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.557639 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.570666 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c16a4e4c-f361-42ef-9af1-e7601b14ca69-pod-info" (OuterVolumeSpecName: "pod-info") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.570764 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-kube-api-access-28d8b" (OuterVolumeSpecName: "kube-api-access-28d8b") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "kube-api-access-28d8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.572089 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.574989 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-config-data" (OuterVolumeSpecName: "config-data") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.608295 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28d8b\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-kube-api-access-28d8b\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.608334 4575 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.608358 4575 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.608367 4575 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.608377 4575 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c16a4e4c-f361-42ef-9af1-e7601b14ca69-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.608385 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.608393 4575 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.608401 4575 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c16a4e4c-f361-42ef-9af1-e7601b14ca69-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.608408 4575 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.653242 4575 generic.go:334] "Generic (PLEG): container finished" podID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" containerID="827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717" exitCode=0 Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.653286 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.653308 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c16a4e4c-f361-42ef-9af1-e7601b14ca69","Type":"ContainerDied","Data":"827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717"} Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.654021 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c16a4e4c-f361-42ef-9af1-e7601b14ca69","Type":"ContainerDied","Data":"76184077f91faaed087ee31a80894f297f5cd645cb59f18b89e775f08eb82455"} Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.654053 4575 scope.go:117] "RemoveContainer" containerID="827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.655896 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-server-conf" (OuterVolumeSpecName: "server-conf") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.656250 4575 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.660710 4575 generic.go:334] "Generic (PLEG): container finished" podID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" containerID="2d02ce72a5f63bcbca2e3120fa4a39e38e4cfa8d01de232b93c3fde2b1e34412" exitCode=0 Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.661662 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5","Type":"ContainerDied","Data":"2d02ce72a5f63bcbca2e3120fa4a39e38e4cfa8d01de232b93c3fde2b1e34412"} Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.664766 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.689356 4575 scope.go:117] "RemoveContainer" containerID="aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.722197 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-confd\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.722349 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-plugins\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.722427 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-pod-info\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.722625 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-tls\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.722810 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-erlang-cookie-secret\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.725137 4575 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c16a4e4c-f361-42ef-9af1-e7601b14ca69-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.725168 4575 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.738716 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.739523 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-pod-info" (OuterVolumeSpecName: "pod-info") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.743251 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.750013 4575 scope.go:117] "RemoveContainer" containerID="827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717" Oct 04 05:01:30 crc kubenswrapper[4575]: E1004 05:01:30.754491 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717\": container with ID starting with 827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717 not found: ID does not exist" containerID="827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.754543 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717"} err="failed to get container status \"827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717\": rpc error: code = NotFound desc = could not find container \"827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717\": container with ID starting with 827e903979325cf069747b101c7fd484ec67dfa222915bf75afc6a70a6406717 not found: ID does not exist" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.754575 4575 scope.go:117] "RemoveContainer" containerID="aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d" Oct 04 05:01:30 crc kubenswrapper[4575]: E1004 05:01:30.755189 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d\": container with ID starting with aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d not found: ID does not exist" containerID="aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.755222 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d"} err="failed to get container status \"aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d\": rpc error: code = NotFound desc = could not find container \"aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d\": container with ID starting with aae47b0aa8e4f4252072acbb966cf7eb0a0bdb105d3258d1c9d5633e2a194f0d not found: ID does not exist" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.773469 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.832982 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-plugins-conf\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.833148 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-server-conf\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.833241 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.833437 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jksjm\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-kube-api-access-jksjm\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.836957 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.839669 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-erlang-cookie\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.839787 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-config-data\") pod \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\" (UID: \"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5\") " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.840656 4575 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.840680 4575 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.840693 4575 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.840734 4575 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.840746 4575 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.841704 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.845535 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-kube-api-access-jksjm" (OuterVolumeSpecName: "kube-api-access-jksjm") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "kube-api-access-jksjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.856165 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.871175 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c16a4e4c-f361-42ef-9af1-e7601b14ca69" (UID: "c16a4e4c-f361-42ef-9af1-e7601b14ca69"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.882497 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-config-data" (OuterVolumeSpecName: "config-data") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.907954 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-server-conf" (OuterVolumeSpecName: "server-conf") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.943492 4575 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.943555 4575 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.943571 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jksjm\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-kube-api-access-jksjm\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.943605 4575 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.943620 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.943632 4575 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c16a4e4c-f361-42ef-9af1-e7601b14ca69-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.950619 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" (UID: "c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:30 crc kubenswrapper[4575]: I1004 05:01:30.971209 4575 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.045618 4575 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.045666 4575 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.080249 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.090097 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.112975 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 05:01:31 crc kubenswrapper[4575]: E1004 05:01:31.113405 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c0652b0-f1a2-4439-87b1-2a083fdbb535" containerName="keystone-cron" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.113426 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c0652b0-f1a2-4439-87b1-2a083fdbb535" containerName="keystone-cron" Oct 04 05:01:31 crc kubenswrapper[4575]: E1004 05:01:31.113441 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" containerName="setup-container" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.113447 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" containerName="setup-container" Oct 04 05:01:31 crc kubenswrapper[4575]: E1004 05:01:31.113460 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" containerName="setup-container" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.113467 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" containerName="setup-container" Oct 04 05:01:31 crc kubenswrapper[4575]: E1004 05:01:31.113479 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" containerName="rabbitmq" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.113485 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" containerName="rabbitmq" Oct 04 05:01:31 crc kubenswrapper[4575]: E1004 05:01:31.113528 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" containerName="rabbitmq" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.113536 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" containerName="rabbitmq" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.113730 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" containerName="rabbitmq" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.113746 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c0652b0-f1a2-4439-87b1-2a083fdbb535" containerName="keystone-cron" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.113774 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" containerName="rabbitmq" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.115257 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.120427 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.120644 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.120966 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.121016 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.121076 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-p9gkh" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.122396 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.130521 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.134065 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.152157 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c02c5e52-debf-4152-8fe1-492d1e88372b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.152396 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c02c5e52-debf-4152-8fe1-492d1e88372b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.152525 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c02c5e52-debf-4152-8fe1-492d1e88372b-config-data\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.152636 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.152708 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.152824 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c02c5e52-debf-4152-8fe1-492d1e88372b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.152947 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmjwp\" (UniqueName: \"kubernetes.io/projected/c02c5e52-debf-4152-8fe1-492d1e88372b-kube-api-access-dmjwp\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.153056 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.153131 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.153397 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c02c5e52-debf-4152-8fe1-492d1e88372b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.153488 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255276 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c02c5e52-debf-4152-8fe1-492d1e88372b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255327 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c02c5e52-debf-4152-8fe1-492d1e88372b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255378 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255397 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c02c5e52-debf-4152-8fe1-492d1e88372b-config-data\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255417 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255439 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c02c5e52-debf-4152-8fe1-492d1e88372b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255504 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmjwp\" (UniqueName: \"kubernetes.io/projected/c02c5e52-debf-4152-8fe1-492d1e88372b-kube-api-access-dmjwp\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255544 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255568 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255632 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c02c5e52-debf-4152-8fe1-492d1e88372b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.255655 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.256149 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.258825 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c02c5e52-debf-4152-8fe1-492d1e88372b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.259254 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.260054 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c02c5e52-debf-4152-8fe1-492d1e88372b-config-data\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.261223 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c02c5e52-debf-4152-8fe1-492d1e88372b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.261377 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.262935 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.263509 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c02c5e52-debf-4152-8fe1-492d1e88372b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.264150 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c02c5e52-debf-4152-8fe1-492d1e88372b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.275810 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c02c5e52-debf-4152-8fe1-492d1e88372b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.278647 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmjwp\" (UniqueName: \"kubernetes.io/projected/c02c5e52-debf-4152-8fe1-492d1e88372b-kube-api-access-dmjwp\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.298030 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c02c5e52-debf-4152-8fe1-492d1e88372b\") " pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.321526 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c16a4e4c-f361-42ef-9af1-e7601b14ca69" path="/var/lib/kubelet/pods/c16a4e4c-f361-42ef-9af1-e7601b14ca69/volumes" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.436885 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.677635 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.677690 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5","Type":"ContainerDied","Data":"f1297539426409d4a22a7171c35b610bbff4b101ef9949a0c3c3b453a625c0d1"} Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.677783 4575 scope.go:117] "RemoveContainer" containerID="2d02ce72a5f63bcbca2e3120fa4a39e38e4cfa8d01de232b93c3fde2b1e34412" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.726497 4575 scope.go:117] "RemoveContainer" containerID="b1fa4225952bfbb53d63c7c3c37d6673d544f5c73aa3fa918767cc4e8a9e89ef" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.729391 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.736235 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.765606 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.773296 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.782979 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.786929 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.788365 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.792240 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.793117 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.793243 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zks2v" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.793315 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.793385 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876172 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876614 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876664 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d699d41f-7e1a-49bc-a88b-1e3906639c7d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876703 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876730 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d699d41f-7e1a-49bc-a88b-1e3906639c7d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876773 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876828 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d699d41f-7e1a-49bc-a88b-1e3906639c7d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876864 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhzkt\" (UniqueName: \"kubernetes.io/projected/d699d41f-7e1a-49bc-a88b-1e3906639c7d-kube-api-access-qhzkt\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876888 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d699d41f-7e1a-49bc-a88b-1e3906639c7d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876912 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d699d41f-7e1a-49bc-a88b-1e3906639c7d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.876945 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.927804 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979329 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979427 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979462 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d699d41f-7e1a-49bc-a88b-1e3906639c7d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979490 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979508 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d699d41f-7e1a-49bc-a88b-1e3906639c7d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979535 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979575 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d699d41f-7e1a-49bc-a88b-1e3906639c7d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979612 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhzkt\" (UniqueName: \"kubernetes.io/projected/d699d41f-7e1a-49bc-a88b-1e3906639c7d-kube-api-access-qhzkt\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979630 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d699d41f-7e1a-49bc-a88b-1e3906639c7d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979654 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d699d41f-7e1a-49bc-a88b-1e3906639c7d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.979686 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.980116 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.981489 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.981783 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.982440 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d699d41f-7e1a-49bc-a88b-1e3906639c7d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.987187 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d699d41f-7e1a-49bc-a88b-1e3906639c7d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.988185 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d699d41f-7e1a-49bc-a88b-1e3906639c7d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.988617 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d699d41f-7e1a-49bc-a88b-1e3906639c7d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.988963 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.993306 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d699d41f-7e1a-49bc-a88b-1e3906639c7d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:31 crc kubenswrapper[4575]: I1004 05:01:31.998434 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d699d41f-7e1a-49bc-a88b-1e3906639c7d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.007401 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhzkt\" (UniqueName: \"kubernetes.io/projected/d699d41f-7e1a-49bc-a88b-1e3906639c7d-kube-api-access-qhzkt\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.012652 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d699d41f-7e1a-49bc-a88b-1e3906639c7d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.118788 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.652272 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.699997 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c02c5e52-debf-4152-8fe1-492d1e88372b","Type":"ContainerStarted","Data":"96e8107bbd69d4a16e32a0d32cbf45368659ae2711a0e918b8626375cc31332d"} Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.703803 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d699d41f-7e1a-49bc-a88b-1e3906639c7d","Type":"ContainerStarted","Data":"dc5a2f4f3d62bf512564365df1d184c200ca0ed7974f2eaf12ad170951d106c2"} Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.902575 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-5dvgl"] Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.904560 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.907845 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.925300 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-5dvgl"] Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.998054 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.998180 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.998230 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.998263 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-svc\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.998335 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-config\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.998446 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqtv9\" (UniqueName: \"kubernetes.io/projected/25676be9-1f97-4050-8b94-0ee513fcd907-kube-api-access-mqtv9\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:32 crc kubenswrapper[4575]: I1004 05:01:32.998497 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.101167 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqtv9\" (UniqueName: \"kubernetes.io/projected/25676be9-1f97-4050-8b94-0ee513fcd907-kube-api-access-mqtv9\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.101297 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.101413 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.101539 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.101616 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.101667 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-svc\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.101780 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-config\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.102914 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-config\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.102984 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.103029 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.103622 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-svc\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.103738 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.104217 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.136214 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqtv9\" (UniqueName: \"kubernetes.io/projected/25676be9-1f97-4050-8b94-0ee513fcd907-kube-api-access-mqtv9\") pod \"dnsmasq-dns-5576978c7c-5dvgl\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.227477 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.325963 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5" path="/var/lib/kubelet/pods/c5dbe6c7-0dd0-45b5-97f0-e3686cb31df5/volumes" Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.633076 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-5dvgl"] Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.723282 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c02c5e52-debf-4152-8fe1-492d1e88372b","Type":"ContainerStarted","Data":"6dfb329502a42d8a52139a20e4021057a916b9dc55d737abff17679613dc0fed"} Oct 04 05:01:33 crc kubenswrapper[4575]: I1004 05:01:33.725255 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" event={"ID":"25676be9-1f97-4050-8b94-0ee513fcd907","Type":"ContainerStarted","Data":"9eda1bee8eaa63aab13670d9f40a3f179fd033072e9f36e290e6a57234b1372b"} Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.692956 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k2m6s"] Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.695543 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.708817 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2m6s"] Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.742456 4575 generic.go:334] "Generic (PLEG): container finished" podID="25676be9-1f97-4050-8b94-0ee513fcd907" containerID="0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872" exitCode=0 Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.742529 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" event={"ID":"25676be9-1f97-4050-8b94-0ee513fcd907","Type":"ContainerDied","Data":"0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872"} Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.746714 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d699d41f-7e1a-49bc-a88b-1e3906639c7d","Type":"ContainerStarted","Data":"26de991664e22db92fd87888a388eda8dc856cc4708fab89464376f00fe01270"} Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.849545 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpzts\" (UniqueName: \"kubernetes.io/projected/63fbf48c-cce8-419b-ab1c-6a82c82914b2-kube-api-access-fpzts\") pod \"redhat-marketplace-k2m6s\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.849917 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-utilities\") pod \"redhat-marketplace-k2m6s\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.850190 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-catalog-content\") pod \"redhat-marketplace-k2m6s\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.951451 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpzts\" (UniqueName: \"kubernetes.io/projected/63fbf48c-cce8-419b-ab1c-6a82c82914b2-kube-api-access-fpzts\") pod \"redhat-marketplace-k2m6s\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.951560 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-utilities\") pod \"redhat-marketplace-k2m6s\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.951702 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-catalog-content\") pod \"redhat-marketplace-k2m6s\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.952255 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-utilities\") pod \"redhat-marketplace-k2m6s\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.952283 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-catalog-content\") pod \"redhat-marketplace-k2m6s\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:34 crc kubenswrapper[4575]: I1004 05:01:34.969441 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpzts\" (UniqueName: \"kubernetes.io/projected/63fbf48c-cce8-419b-ab1c-6a82c82914b2-kube-api-access-fpzts\") pod \"redhat-marketplace-k2m6s\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.014856 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.311110 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:01:35 crc kubenswrapper[4575]: E1004 05:01:35.311942 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:01:35 crc kubenswrapper[4575]: W1004 05:01:35.477741 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63fbf48c_cce8_419b_ab1c_6a82c82914b2.slice/crio-bdfd664843326e38df7872676555966caa3c16922c66cbfe3d8ef77355722f95 WatchSource:0}: Error finding container bdfd664843326e38df7872676555966caa3c16922c66cbfe3d8ef77355722f95: Status 404 returned error can't find the container with id bdfd664843326e38df7872676555966caa3c16922c66cbfe3d8ef77355722f95 Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.478817 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2m6s"] Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.557576 4575 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-b86b879b4-tlf42" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.558102 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.772904 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" event={"ID":"25676be9-1f97-4050-8b94-0ee513fcd907","Type":"ContainerStarted","Data":"4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357"} Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.774629 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.777053 4575 generic.go:334] "Generic (PLEG): container finished" podID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerID="f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451" exitCode=0 Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.777230 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2m6s" event={"ID":"63fbf48c-cce8-419b-ab1c-6a82c82914b2","Type":"ContainerDied","Data":"f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451"} Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.777295 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2m6s" event={"ID":"63fbf48c-cce8-419b-ab1c-6a82c82914b2","Type":"ContainerStarted","Data":"bdfd664843326e38df7872676555966caa3c16922c66cbfe3d8ef77355722f95"} Oct 04 05:01:35 crc kubenswrapper[4575]: I1004 05:01:35.810046 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" podStartSLOduration=3.81002783 podStartE2EDuration="3.81002783s" podCreationTimestamp="2025-10-04 05:01:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:01:35.803695386 +0000 UTC m=+1647.132254340" watchObservedRunningTime="2025-10-04 05:01:35.81002783 +0000 UTC m=+1647.138586644" Oct 04 05:01:37 crc kubenswrapper[4575]: I1004 05:01:37.807978 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2m6s" event={"ID":"63fbf48c-cce8-419b-ab1c-6a82c82914b2","Type":"ContainerStarted","Data":"06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6"} Oct 04 05:01:38 crc kubenswrapper[4575]: I1004 05:01:38.817508 4575 generic.go:334] "Generic (PLEG): container finished" podID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerID="06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6" exitCode=0 Oct 04 05:01:38 crc kubenswrapper[4575]: I1004 05:01:38.817577 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2m6s" event={"ID":"63fbf48c-cce8-419b-ab1c-6a82c82914b2","Type":"ContainerDied","Data":"06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6"} Oct 04 05:01:39 crc kubenswrapper[4575]: I1004 05:01:39.828903 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2m6s" event={"ID":"63fbf48c-cce8-419b-ab1c-6a82c82914b2","Type":"ContainerStarted","Data":"251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53"} Oct 04 05:01:39 crc kubenswrapper[4575]: I1004 05:01:39.854621 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k2m6s" podStartSLOduration=2.38085068 podStartE2EDuration="5.854583567s" podCreationTimestamp="2025-10-04 05:01:34 +0000 UTC" firstStartedPulling="2025-10-04 05:01:35.779431871 +0000 UTC m=+1647.107990685" lastFinishedPulling="2025-10-04 05:01:39.253164768 +0000 UTC m=+1650.581723572" observedRunningTime="2025-10-04 05:01:39.845203524 +0000 UTC m=+1651.173762338" watchObservedRunningTime="2025-10-04 05:01:39.854583567 +0000 UTC m=+1651.183142391" Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.803017 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.847347 4575 generic.go:334] "Generic (PLEG): container finished" podID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerID="bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149" exitCode=137 Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.847413 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b86b879b4-tlf42" Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.847484 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerDied","Data":"bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149"} Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.847516 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b86b879b4-tlf42" event={"ID":"115cafce-0f37-4b82-8b55-c2bbf7297226","Type":"ContainerDied","Data":"7263cde13925ae3c7d67b8bc6180bda1accc8f2ced8fa93c3aea0d0211094759"} Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.847536 4575 scope.go:117] "RemoveContainer" containerID="ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b" Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.969384 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-secret-key\") pod \"115cafce-0f37-4b82-8b55-c2bbf7297226\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.969543 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/115cafce-0f37-4b82-8b55-c2bbf7297226-logs\") pod \"115cafce-0f37-4b82-8b55-c2bbf7297226\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.969689 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-scripts\") pod \"115cafce-0f37-4b82-8b55-c2bbf7297226\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.969721 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-combined-ca-bundle\") pod \"115cafce-0f37-4b82-8b55-c2bbf7297226\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.970185 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/115cafce-0f37-4b82-8b55-c2bbf7297226-logs" (OuterVolumeSpecName: "logs") pod "115cafce-0f37-4b82-8b55-c2bbf7297226" (UID: "115cafce-0f37-4b82-8b55-c2bbf7297226"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.970644 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-config-data\") pod \"115cafce-0f37-4b82-8b55-c2bbf7297226\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.970699 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4rf6\" (UniqueName: \"kubernetes.io/projected/115cafce-0f37-4b82-8b55-c2bbf7297226-kube-api-access-v4rf6\") pod \"115cafce-0f37-4b82-8b55-c2bbf7297226\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.970787 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-tls-certs\") pod \"115cafce-0f37-4b82-8b55-c2bbf7297226\" (UID: \"115cafce-0f37-4b82-8b55-c2bbf7297226\") " Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.971457 4575 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/115cafce-0f37-4b82-8b55-c2bbf7297226-logs\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.976068 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/115cafce-0f37-4b82-8b55-c2bbf7297226-kube-api-access-v4rf6" (OuterVolumeSpecName: "kube-api-access-v4rf6") pod "115cafce-0f37-4b82-8b55-c2bbf7297226" (UID: "115cafce-0f37-4b82-8b55-c2bbf7297226"). InnerVolumeSpecName "kube-api-access-v4rf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.990854 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "115cafce-0f37-4b82-8b55-c2bbf7297226" (UID: "115cafce-0f37-4b82-8b55-c2bbf7297226"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:40 crc kubenswrapper[4575]: I1004 05:01:40.998408 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-scripts" (OuterVolumeSpecName: "scripts") pod "115cafce-0f37-4b82-8b55-c2bbf7297226" (UID: "115cafce-0f37-4b82-8b55-c2bbf7297226"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.007279 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-config-data" (OuterVolumeSpecName: "config-data") pod "115cafce-0f37-4b82-8b55-c2bbf7297226" (UID: "115cafce-0f37-4b82-8b55-c2bbf7297226"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.008291 4575 scope.go:117] "RemoveContainer" containerID="bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.008303 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "115cafce-0f37-4b82-8b55-c2bbf7297226" (UID: "115cafce-0f37-4b82-8b55-c2bbf7297226"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.046580 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "115cafce-0f37-4b82-8b55-c2bbf7297226" (UID: "115cafce-0f37-4b82-8b55-c2bbf7297226"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.073738 4575 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.073788 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.073806 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/115cafce-0f37-4b82-8b55-c2bbf7297226-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.073821 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4rf6\" (UniqueName: \"kubernetes.io/projected/115cafce-0f37-4b82-8b55-c2bbf7297226-kube-api-access-v4rf6\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.073838 4575 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.073850 4575 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/115cafce-0f37-4b82-8b55-c2bbf7297226-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.129367 4575 scope.go:117] "RemoveContainer" containerID="ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b" Oct 04 05:01:41 crc kubenswrapper[4575]: E1004 05:01:41.129927 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b\": container with ID starting with ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b not found: ID does not exist" containerID="ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.129986 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b"} err="failed to get container status \"ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b\": rpc error: code = NotFound desc = could not find container \"ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b\": container with ID starting with ec31a3c72e536c75599b8b39f9fee9a0111e3891d88bfe71388d86afebe8c20b not found: ID does not exist" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.130021 4575 scope.go:117] "RemoveContainer" containerID="bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149" Oct 04 05:01:41 crc kubenswrapper[4575]: E1004 05:01:41.130508 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149\": container with ID starting with bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149 not found: ID does not exist" containerID="bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.130552 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149"} err="failed to get container status \"bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149\": rpc error: code = NotFound desc = could not find container \"bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149\": container with ID starting with bf4c2981050327e08c1a6acfb4a3ce3343505ee719e080296bdc51c746038149 not found: ID does not exist" Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.183030 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b86b879b4-tlf42"] Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.191442 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-b86b879b4-tlf42"] Oct 04 05:01:41 crc kubenswrapper[4575]: I1004 05:01:41.321707 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" path="/var/lib/kubelet/pods/115cafce-0f37-4b82-8b55-c2bbf7297226/volumes" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.228659 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.298735 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-cx2ll"] Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.299048 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" podUID="ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" containerName="dnsmasq-dns" containerID="cri-o://84f729f8e5b939d7e4745c7ca6e459581f6de204fe07d1cce6aed2995b8b7ba1" gracePeriod=10 Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.542221 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-667c9c995c-wgqqr"] Oct 04 05:01:43 crc kubenswrapper[4575]: E1004 05:01:43.550185 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550230 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: E1004 05:01:43.550258 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550267 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: E1004 05:01:43.550284 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550291 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: E1004 05:01:43.550314 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon-log" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550322 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon-log" Oct 04 05:01:43 crc kubenswrapper[4575]: E1004 05:01:43.550333 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550340 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550562 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550615 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550637 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550653 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550667 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550684 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon-log" Oct 04 05:01:43 crc kubenswrapper[4575]: E1004 05:01:43.550903 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.550913 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="115cafce-0f37-4b82-8b55-c2bbf7297226" containerName="horizon" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.551815 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.589726 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667c9c995c-wgqqr"] Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.643833 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-dns-svc\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.643959 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-config\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.644040 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brc2g\" (UniqueName: \"kubernetes.io/projected/a830ac92-f57f-4f4d-85e6-55de886289bb-kube-api-access-brc2g\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.644297 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-dns-swift-storage-0\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.644413 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-ovsdbserver-sb\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.644460 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-ovsdbserver-nb\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.644502 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-openstack-edpm-ipam\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.759928 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-config\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.763671 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brc2g\" (UniqueName: \"kubernetes.io/projected/a830ac92-f57f-4f4d-85e6-55de886289bb-kube-api-access-brc2g\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.764362 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-dns-swift-storage-0\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.765148 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-ovsdbserver-sb\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.765540 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-ovsdbserver-nb\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.765813 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-openstack-edpm-ipam\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.766158 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-dns-svc\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.767204 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-dns-svc\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.770363 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-dns-swift-storage-0\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.761945 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-config\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.771221 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-ovsdbserver-nb\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.771803 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-ovsdbserver-sb\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.772532 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a830ac92-f57f-4f4d-85e6-55de886289bb-openstack-edpm-ipam\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.826237 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brc2g\" (UniqueName: \"kubernetes.io/projected/a830ac92-f57f-4f4d-85e6-55de886289bb-kube-api-access-brc2g\") pod \"dnsmasq-dns-667c9c995c-wgqqr\" (UID: \"a830ac92-f57f-4f4d-85e6-55de886289bb\") " pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.887235 4575 generic.go:334] "Generic (PLEG): container finished" podID="ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" containerID="84f729f8e5b939d7e4745c7ca6e459581f6de204fe07d1cce6aed2995b8b7ba1" exitCode=0 Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.887279 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" event={"ID":"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443","Type":"ContainerDied","Data":"84f729f8e5b939d7e4745c7ca6e459581f6de204fe07d1cce6aed2995b8b7ba1"} Oct 04 05:01:43 crc kubenswrapper[4575]: I1004 05:01:43.894396 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.082367 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.175684 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-nb\") pod \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.175749 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-swift-storage-0\") pod \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.175859 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-svc\") pod \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.175957 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlxql\" (UniqueName: \"kubernetes.io/projected/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-kube-api-access-zlxql\") pod \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.175980 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-sb\") pod \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.176015 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-config\") pod \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\" (UID: \"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443\") " Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.181228 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-kube-api-access-zlxql" (OuterVolumeSpecName: "kube-api-access-zlxql") pod "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" (UID: "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443"). InnerVolumeSpecName "kube-api-access-zlxql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.250394 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-config" (OuterVolumeSpecName: "config") pod "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" (UID: "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.269956 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" (UID: "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.272013 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" (UID: "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.276486 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" (UID: "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.278336 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlxql\" (UniqueName: \"kubernetes.io/projected/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-kube-api-access-zlxql\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.278437 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-config\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.278508 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.278769 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.278885 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.296436 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" (UID: "ff79a71f-d22f-4fb7-8c67-4c9f1d97a443"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.381260 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.554877 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667c9c995c-wgqqr"] Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.904035 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.904053 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-cx2ll" event={"ID":"ff79a71f-d22f-4fb7-8c67-4c9f1d97a443","Type":"ContainerDied","Data":"e3bfef8b0c4bc46bbcc465711ecf12a02e3cb5445bd552bda3b91e5591a8fc86"} Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.904723 4575 scope.go:117] "RemoveContainer" containerID="84f729f8e5b939d7e4745c7ca6e459581f6de204fe07d1cce6aed2995b8b7ba1" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.910785 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" event={"ID":"a830ac92-f57f-4f4d-85e6-55de886289bb","Type":"ContainerStarted","Data":"5c3677ba8371f57285c2afe6f1454fbab7caaeff94caf21b40b8e45db22f9458"} Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.910819 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" event={"ID":"a830ac92-f57f-4f4d-85e6-55de886289bb","Type":"ContainerStarted","Data":"074df4b7e1b2e13be70410ff2afe823bf88c8205c15ab8ef7f43d25e420237c6"} Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.942394 4575 scope.go:117] "RemoveContainer" containerID="3e6c1087351b9cdd29bc0a5fea0e2388643d779e44f47149978f74bd53089d36" Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.960929 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-cx2ll"] Oct 04 05:01:44 crc kubenswrapper[4575]: I1004 05:01:44.970518 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-cx2ll"] Oct 04 05:01:45 crc kubenswrapper[4575]: I1004 05:01:45.015933 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:45 crc kubenswrapper[4575]: I1004 05:01:45.017490 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:45 crc kubenswrapper[4575]: I1004 05:01:45.067472 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:45 crc kubenswrapper[4575]: I1004 05:01:45.326713 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" path="/var/lib/kubelet/pods/ff79a71f-d22f-4fb7-8c67-4c9f1d97a443/volumes" Oct 04 05:01:45 crc kubenswrapper[4575]: I1004 05:01:45.923436 4575 generic.go:334] "Generic (PLEG): container finished" podID="a830ac92-f57f-4f4d-85e6-55de886289bb" containerID="5c3677ba8371f57285c2afe6f1454fbab7caaeff94caf21b40b8e45db22f9458" exitCode=0 Oct 04 05:01:45 crc kubenswrapper[4575]: I1004 05:01:45.923683 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" event={"ID":"a830ac92-f57f-4f4d-85e6-55de886289bb","Type":"ContainerDied","Data":"5c3677ba8371f57285c2afe6f1454fbab7caaeff94caf21b40b8e45db22f9458"} Oct 04 05:01:45 crc kubenswrapper[4575]: I1004 05:01:45.994189 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:46 crc kubenswrapper[4575]: I1004 05:01:46.055221 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2m6s"] Oct 04 05:01:46 crc kubenswrapper[4575]: I1004 05:01:46.953327 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" event={"ID":"a830ac92-f57f-4f4d-85e6-55de886289bb","Type":"ContainerStarted","Data":"13e423e3d5b6d59e2264c25922b1f7109c0cf032596e68ba317951a22847a525"} Oct 04 05:01:46 crc kubenswrapper[4575]: I1004 05:01:46.980677 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" podStartSLOduration=3.980654988 podStartE2EDuration="3.980654988s" podCreationTimestamp="2025-10-04 05:01:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:01:46.972412478 +0000 UTC m=+1658.300971302" watchObservedRunningTime="2025-10-04 05:01:46.980654988 +0000 UTC m=+1658.309213802" Oct 04 05:01:47 crc kubenswrapper[4575]: I1004 05:01:47.962929 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:47 crc kubenswrapper[4575]: I1004 05:01:47.963479 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k2m6s" podUID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerName="registry-server" containerID="cri-o://251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53" gracePeriod=2 Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.309812 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:01:48 crc kubenswrapper[4575]: E1004 05:01:48.310616 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.456905 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.578092 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-utilities\") pod \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.578314 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-catalog-content\") pod \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.578394 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpzts\" (UniqueName: \"kubernetes.io/projected/63fbf48c-cce8-419b-ab1c-6a82c82914b2-kube-api-access-fpzts\") pod \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\" (UID: \"63fbf48c-cce8-419b-ab1c-6a82c82914b2\") " Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.579163 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-utilities" (OuterVolumeSpecName: "utilities") pod "63fbf48c-cce8-419b-ab1c-6a82c82914b2" (UID: "63fbf48c-cce8-419b-ab1c-6a82c82914b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.586897 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63fbf48c-cce8-419b-ab1c-6a82c82914b2-kube-api-access-fpzts" (OuterVolumeSpecName: "kube-api-access-fpzts") pod "63fbf48c-cce8-419b-ab1c-6a82c82914b2" (UID: "63fbf48c-cce8-419b-ab1c-6a82c82914b2"). InnerVolumeSpecName "kube-api-access-fpzts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.593614 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63fbf48c-cce8-419b-ab1c-6a82c82914b2" (UID: "63fbf48c-cce8-419b-ab1c-6a82c82914b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.680737 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.680786 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpzts\" (UniqueName: \"kubernetes.io/projected/63fbf48c-cce8-419b-ab1c-6a82c82914b2-kube-api-access-fpzts\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.680801 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fbf48c-cce8-419b-ab1c-6a82c82914b2-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.975348 4575 generic.go:334] "Generic (PLEG): container finished" podID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerID="251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53" exitCode=0 Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.975396 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2m6s" event={"ID":"63fbf48c-cce8-419b-ab1c-6a82c82914b2","Type":"ContainerDied","Data":"251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53"} Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.975456 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2m6s" event={"ID":"63fbf48c-cce8-419b-ab1c-6a82c82914b2","Type":"ContainerDied","Data":"bdfd664843326e38df7872676555966caa3c16922c66cbfe3d8ef77355722f95"} Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.975477 4575 scope.go:117] "RemoveContainer" containerID="251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.976546 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2m6s" Oct 04 05:01:48 crc kubenswrapper[4575]: I1004 05:01:48.994765 4575 scope.go:117] "RemoveContainer" containerID="06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6" Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.028211 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2m6s"] Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.036171 4575 scope.go:117] "RemoveContainer" containerID="f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451" Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.044544 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2m6s"] Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.068832 4575 scope.go:117] "RemoveContainer" containerID="251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53" Oct 04 05:01:49 crc kubenswrapper[4575]: E1004 05:01:49.069851 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53\": container with ID starting with 251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53 not found: ID does not exist" containerID="251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53" Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.069925 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53"} err="failed to get container status \"251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53\": rpc error: code = NotFound desc = could not find container \"251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53\": container with ID starting with 251ec70c281ef71d7f1ef1cc7d0a83471a225765c3551376ba95327d83580f53 not found: ID does not exist" Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.069951 4575 scope.go:117] "RemoveContainer" containerID="06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6" Oct 04 05:01:49 crc kubenswrapper[4575]: E1004 05:01:49.070248 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6\": container with ID starting with 06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6 not found: ID does not exist" containerID="06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6" Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.070344 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6"} err="failed to get container status \"06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6\": rpc error: code = NotFound desc = could not find container \"06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6\": container with ID starting with 06ee9c4b2f50c5f4a404047cea3748bbc8751d51a0143e8ff27a2e75f36eb4b6 not found: ID does not exist" Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.070425 4575 scope.go:117] "RemoveContainer" containerID="f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451" Oct 04 05:01:49 crc kubenswrapper[4575]: E1004 05:01:49.070889 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451\": container with ID starting with f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451 not found: ID does not exist" containerID="f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451" Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.070917 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451"} err="failed to get container status \"f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451\": rpc error: code = NotFound desc = could not find container \"f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451\": container with ID starting with f1796143d09cedb2e7c4dc981ec85dd29790d14e8fb4bf9c6d1cfed7bb763451 not found: ID does not exist" Oct 04 05:01:49 crc kubenswrapper[4575]: I1004 05:01:49.329742 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" path="/var/lib/kubelet/pods/63fbf48c-cce8-419b-ab1c-6a82c82914b2/volumes" Oct 04 05:01:53 crc kubenswrapper[4575]: I1004 05:01:53.896748 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-667c9c995c-wgqqr" Oct 04 05:01:53 crc kubenswrapper[4575]: I1004 05:01:53.976268 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-5dvgl"] Oct 04 05:01:53 crc kubenswrapper[4575]: I1004 05:01:53.976781 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" podUID="25676be9-1f97-4050-8b94-0ee513fcd907" containerName="dnsmasq-dns" containerID="cri-o://4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357" gracePeriod=10 Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.546075 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.702650 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-svc\") pod \"25676be9-1f97-4050-8b94-0ee513fcd907\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.702833 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqtv9\" (UniqueName: \"kubernetes.io/projected/25676be9-1f97-4050-8b94-0ee513fcd907-kube-api-access-mqtv9\") pod \"25676be9-1f97-4050-8b94-0ee513fcd907\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.702940 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-config\") pod \"25676be9-1f97-4050-8b94-0ee513fcd907\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.702979 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-nb\") pod \"25676be9-1f97-4050-8b94-0ee513fcd907\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.703065 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-openstack-edpm-ipam\") pod \"25676be9-1f97-4050-8b94-0ee513fcd907\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.703089 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-sb\") pod \"25676be9-1f97-4050-8b94-0ee513fcd907\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.703137 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-swift-storage-0\") pod \"25676be9-1f97-4050-8b94-0ee513fcd907\" (UID: \"25676be9-1f97-4050-8b94-0ee513fcd907\") " Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.710055 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25676be9-1f97-4050-8b94-0ee513fcd907-kube-api-access-mqtv9" (OuterVolumeSpecName: "kube-api-access-mqtv9") pod "25676be9-1f97-4050-8b94-0ee513fcd907" (UID: "25676be9-1f97-4050-8b94-0ee513fcd907"). InnerVolumeSpecName "kube-api-access-mqtv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.762415 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "25676be9-1f97-4050-8b94-0ee513fcd907" (UID: "25676be9-1f97-4050-8b94-0ee513fcd907"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.775493 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "25676be9-1f97-4050-8b94-0ee513fcd907" (UID: "25676be9-1f97-4050-8b94-0ee513fcd907"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.778868 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-config" (OuterVolumeSpecName: "config") pod "25676be9-1f97-4050-8b94-0ee513fcd907" (UID: "25676be9-1f97-4050-8b94-0ee513fcd907"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.780420 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "25676be9-1f97-4050-8b94-0ee513fcd907" (UID: "25676be9-1f97-4050-8b94-0ee513fcd907"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.781571 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "25676be9-1f97-4050-8b94-0ee513fcd907" (UID: "25676be9-1f97-4050-8b94-0ee513fcd907"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.787040 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "25676be9-1f97-4050-8b94-0ee513fcd907" (UID: "25676be9-1f97-4050-8b94-0ee513fcd907"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.805317 4575 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-config\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.805358 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.805375 4575 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.805389 4575 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.805403 4575 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.805415 4575 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25676be9-1f97-4050-8b94-0ee513fcd907-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:54 crc kubenswrapper[4575]: I1004 05:01:54.805426 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqtv9\" (UniqueName: \"kubernetes.io/projected/25676be9-1f97-4050-8b94-0ee513fcd907-kube-api-access-mqtv9\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.042806 4575 generic.go:334] "Generic (PLEG): container finished" podID="25676be9-1f97-4050-8b94-0ee513fcd907" containerID="4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357" exitCode=0 Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.042909 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" event={"ID":"25676be9-1f97-4050-8b94-0ee513fcd907","Type":"ContainerDied","Data":"4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357"} Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.042964 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" event={"ID":"25676be9-1f97-4050-8b94-0ee513fcd907","Type":"ContainerDied","Data":"9eda1bee8eaa63aab13670d9f40a3f179fd033072e9f36e290e6a57234b1372b"} Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.042986 4575 scope.go:117] "RemoveContainer" containerID="4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357" Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.043806 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-5dvgl" Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.086888 4575 scope.go:117] "RemoveContainer" containerID="0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872" Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.099978 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-5dvgl"] Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.117117 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-5dvgl"] Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.151662 4575 scope.go:117] "RemoveContainer" containerID="4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357" Oct 04 05:01:55 crc kubenswrapper[4575]: E1004 05:01:55.152074 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357\": container with ID starting with 4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357 not found: ID does not exist" containerID="4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357" Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.152191 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357"} err="failed to get container status \"4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357\": rpc error: code = NotFound desc = could not find container \"4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357\": container with ID starting with 4c9fc05cd93f6fbace47c23a52da62a242d4341b4570f929ffa742947a153357 not found: ID does not exist" Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.152289 4575 scope.go:117] "RemoveContainer" containerID="0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872" Oct 04 05:01:55 crc kubenswrapper[4575]: E1004 05:01:55.152842 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872\": container with ID starting with 0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872 not found: ID does not exist" containerID="0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872" Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.152890 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872"} err="failed to get container status \"0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872\": rpc error: code = NotFound desc = could not find container \"0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872\": container with ID starting with 0b2662f1ce0af29b25e6170ee9aee14f06b13b315bf14a65f33924e54fd0b872 not found: ID does not exist" Oct 04 05:01:55 crc kubenswrapper[4575]: I1004 05:01:55.322015 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25676be9-1f97-4050-8b94-0ee513fcd907" path="/var/lib/kubelet/pods/25676be9-1f97-4050-8b94-0ee513fcd907/volumes" Oct 04 05:02:02 crc kubenswrapper[4575]: I1004 05:02:02.309907 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:02:02 crc kubenswrapper[4575]: E1004 05:02:02.310654 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:02:06 crc kubenswrapper[4575]: I1004 05:02:06.210450 4575 generic.go:334] "Generic (PLEG): container finished" podID="c02c5e52-debf-4152-8fe1-492d1e88372b" containerID="6dfb329502a42d8a52139a20e4021057a916b9dc55d737abff17679613dc0fed" exitCode=0 Oct 04 05:02:06 crc kubenswrapper[4575]: I1004 05:02:06.210522 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c02c5e52-debf-4152-8fe1-492d1e88372b","Type":"ContainerDied","Data":"6dfb329502a42d8a52139a20e4021057a916b9dc55d737abff17679613dc0fed"} Oct 04 05:02:07 crc kubenswrapper[4575]: I1004 05:02:07.228614 4575 generic.go:334] "Generic (PLEG): container finished" podID="d699d41f-7e1a-49bc-a88b-1e3906639c7d" containerID="26de991664e22db92fd87888a388eda8dc856cc4708fab89464376f00fe01270" exitCode=0 Oct 04 05:02:07 crc kubenswrapper[4575]: I1004 05:02:07.228699 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d699d41f-7e1a-49bc-a88b-1e3906639c7d","Type":"ContainerDied","Data":"26de991664e22db92fd87888a388eda8dc856cc4708fab89464376f00fe01270"} Oct 04 05:02:07 crc kubenswrapper[4575]: I1004 05:02:07.232538 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c02c5e52-debf-4152-8fe1-492d1e88372b","Type":"ContainerStarted","Data":"78a243dd36e62718324c95d5c15783a48ab93606de2c8b6b27fe5ca6158fdc20"} Oct 04 05:02:07 crc kubenswrapper[4575]: I1004 05:02:07.232848 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 05:02:07 crc kubenswrapper[4575]: I1004 05:02:07.311450 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.311425775000004 podStartE2EDuration="36.311425775s" podCreationTimestamp="2025-10-04 05:01:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:02:07.292490435 +0000 UTC m=+1678.621049269" watchObservedRunningTime="2025-10-04 05:02:07.311425775 +0000 UTC m=+1678.639984589" Oct 04 05:02:08 crc kubenswrapper[4575]: I1004 05:02:08.244149 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d699d41f-7e1a-49bc-a88b-1e3906639c7d","Type":"ContainerStarted","Data":"fa7a2822116d85f384dd9b4efe454c94080e8ec3adde567252a235c82f287754"} Oct 04 05:02:08 crc kubenswrapper[4575]: I1004 05:02:08.244925 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.979392 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.979369046 podStartE2EDuration="41.979369046s" podCreationTimestamp="2025-10-04 05:01:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:02:08.277051302 +0000 UTC m=+1679.605610126" watchObservedRunningTime="2025-10-04 05:02:12.979369046 +0000 UTC m=+1684.307927860" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.984854 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87"] Oct 04 05:02:12 crc kubenswrapper[4575]: E1004 05:02:12.985276 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" containerName="dnsmasq-dns" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985295 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" containerName="dnsmasq-dns" Oct 04 05:02:12 crc kubenswrapper[4575]: E1004 05:02:12.985308 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerName="registry-server" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985315 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerName="registry-server" Oct 04 05:02:12 crc kubenswrapper[4575]: E1004 05:02:12.985328 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25676be9-1f97-4050-8b94-0ee513fcd907" containerName="init" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985336 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="25676be9-1f97-4050-8b94-0ee513fcd907" containerName="init" Oct 04 05:02:12 crc kubenswrapper[4575]: E1004 05:02:12.985344 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25676be9-1f97-4050-8b94-0ee513fcd907" containerName="dnsmasq-dns" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985350 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="25676be9-1f97-4050-8b94-0ee513fcd907" containerName="dnsmasq-dns" Oct 04 05:02:12 crc kubenswrapper[4575]: E1004 05:02:12.985374 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" containerName="init" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985381 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" containerName="init" Oct 04 05:02:12 crc kubenswrapper[4575]: E1004 05:02:12.985400 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerName="extract-utilities" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985408 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerName="extract-utilities" Oct 04 05:02:12 crc kubenswrapper[4575]: E1004 05:02:12.985425 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerName="extract-content" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985442 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerName="extract-content" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985640 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff79a71f-d22f-4fb7-8c67-4c9f1d97a443" containerName="dnsmasq-dns" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985650 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="25676be9-1f97-4050-8b94-0ee513fcd907" containerName="dnsmasq-dns" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.985674 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="63fbf48c-cce8-419b-ab1c-6a82c82914b2" containerName="registry-server" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.986280 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.990166 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.990378 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.991391 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.991646 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:02:12 crc kubenswrapper[4575]: I1004 05:02:12.998449 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87"] Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.092771 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.092897 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbgz5\" (UniqueName: \"kubernetes.io/projected/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-kube-api-access-rbgz5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.092938 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.092998 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.194392 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbgz5\" (UniqueName: \"kubernetes.io/projected/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-kube-api-access-rbgz5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.194444 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.194517 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.194565 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.200759 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.201342 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.209241 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.222117 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbgz5\" (UniqueName: \"kubernetes.io/projected/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-kube-api-access-rbgz5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-92l87\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:13 crc kubenswrapper[4575]: I1004 05:02:13.308050 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:14 crc kubenswrapper[4575]: I1004 05:02:14.740332 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87"] Oct 04 05:02:15 crc kubenswrapper[4575]: I1004 05:02:15.309635 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:02:15 crc kubenswrapper[4575]: E1004 05:02:15.310169 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:02:15 crc kubenswrapper[4575]: I1004 05:02:15.325958 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" event={"ID":"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7","Type":"ContainerStarted","Data":"e6eacbd8347edd176b6100c0909987fb03ba53049e0e10ca58ddbe08d9f6311f"} Oct 04 05:02:21 crc kubenswrapper[4575]: I1004 05:02:21.439779 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 05:02:22 crc kubenswrapper[4575]: I1004 05:02:22.127766 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 05:02:27 crc kubenswrapper[4575]: I1004 05:02:27.477848 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" event={"ID":"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7","Type":"ContainerStarted","Data":"c94c6e8f28a3ba5d5a9818ac62d2b448d22779b782fd0825e1cd75482b014119"} Oct 04 05:02:27 crc kubenswrapper[4575]: I1004 05:02:27.506410 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" podStartSLOduration=3.490567664 podStartE2EDuration="15.506389883s" podCreationTimestamp="2025-10-04 05:02:12 +0000 UTC" firstStartedPulling="2025-10-04 05:02:14.743980262 +0000 UTC m=+1686.072539076" lastFinishedPulling="2025-10-04 05:02:26.759802481 +0000 UTC m=+1698.088361295" observedRunningTime="2025-10-04 05:02:27.498282708 +0000 UTC m=+1698.826841522" watchObservedRunningTime="2025-10-04 05:02:27.506389883 +0000 UTC m=+1698.834948687" Oct 04 05:02:28 crc kubenswrapper[4575]: I1004 05:02:28.046068 4575 scope.go:117] "RemoveContainer" containerID="63ea1872a641b4a425e2746229cfdec26f0609d439e87d391c014c1d4a2355b0" Oct 04 05:02:28 crc kubenswrapper[4575]: I1004 05:02:28.310547 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:02:28 crc kubenswrapper[4575]: E1004 05:02:28.310827 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:02:39 crc kubenswrapper[4575]: I1004 05:02:39.348185 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:02:39 crc kubenswrapper[4575]: E1004 05:02:39.349330 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:02:43 crc kubenswrapper[4575]: I1004 05:02:43.622972 4575 generic.go:334] "Generic (PLEG): container finished" podID="727a9553-c9b7-4dab-93fb-c8c9ab1b83a7" containerID="c94c6e8f28a3ba5d5a9818ac62d2b448d22779b782fd0825e1cd75482b014119" exitCode=0 Oct 04 05:02:43 crc kubenswrapper[4575]: I1004 05:02:43.623107 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" event={"ID":"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7","Type":"ContainerDied","Data":"c94c6e8f28a3ba5d5a9818ac62d2b448d22779b782fd0825e1cd75482b014119"} Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.077765 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.259252 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-repo-setup-combined-ca-bundle\") pod \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.259448 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-ssh-key\") pod \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.259610 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbgz5\" (UniqueName: \"kubernetes.io/projected/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-kube-api-access-rbgz5\") pod \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.259750 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-inventory\") pod \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\" (UID: \"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7\") " Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.265974 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-kube-api-access-rbgz5" (OuterVolumeSpecName: "kube-api-access-rbgz5") pod "727a9553-c9b7-4dab-93fb-c8c9ab1b83a7" (UID: "727a9553-c9b7-4dab-93fb-c8c9ab1b83a7"). InnerVolumeSpecName "kube-api-access-rbgz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.266144 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "727a9553-c9b7-4dab-93fb-c8c9ab1b83a7" (UID: "727a9553-c9b7-4dab-93fb-c8c9ab1b83a7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.299213 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "727a9553-c9b7-4dab-93fb-c8c9ab1b83a7" (UID: "727a9553-c9b7-4dab-93fb-c8c9ab1b83a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.301075 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-inventory" (OuterVolumeSpecName: "inventory") pod "727a9553-c9b7-4dab-93fb-c8c9ab1b83a7" (UID: "727a9553-c9b7-4dab-93fb-c8c9ab1b83a7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.362417 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.362463 4575 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.362475 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.362484 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbgz5\" (UniqueName: \"kubernetes.io/projected/727a9553-c9b7-4dab-93fb-c8c9ab1b83a7-kube-api-access-rbgz5\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.650894 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" event={"ID":"727a9553-c9b7-4dab-93fb-c8c9ab1b83a7","Type":"ContainerDied","Data":"e6eacbd8347edd176b6100c0909987fb03ba53049e0e10ca58ddbe08d9f6311f"} Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.650942 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6eacbd8347edd176b6100c0909987fb03ba53049e0e10ca58ddbe08d9f6311f" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.651024 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-92l87" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.748909 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt"] Oct 04 05:02:45 crc kubenswrapper[4575]: E1004 05:02:45.749405 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727a9553-c9b7-4dab-93fb-c8c9ab1b83a7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.749426 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="727a9553-c9b7-4dab-93fb-c8c9ab1b83a7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.749702 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="727a9553-c9b7-4dab-93fb-c8c9ab1b83a7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.750484 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.753282 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.753501 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.760105 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.760528 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.776383 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt"] Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.870933 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-25bjt\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.871365 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-25bjt\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.871630 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jswcd\" (UniqueName: \"kubernetes.io/projected/7a3a20d3-21b8-456a-b46d-bfe264defc14-kube-api-access-jswcd\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-25bjt\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.973498 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jswcd\" (UniqueName: \"kubernetes.io/projected/7a3a20d3-21b8-456a-b46d-bfe264defc14-kube-api-access-jswcd\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-25bjt\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.973668 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-25bjt\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.973728 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-25bjt\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.980603 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-25bjt\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.987333 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-25bjt\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:45 crc kubenswrapper[4575]: I1004 05:02:45.991710 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jswcd\" (UniqueName: \"kubernetes.io/projected/7a3a20d3-21b8-456a-b46d-bfe264defc14-kube-api-access-jswcd\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-25bjt\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:46 crc kubenswrapper[4575]: I1004 05:02:46.070161 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:46 crc kubenswrapper[4575]: I1004 05:02:46.621056 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt"] Oct 04 05:02:46 crc kubenswrapper[4575]: I1004 05:02:46.663502 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" event={"ID":"7a3a20d3-21b8-456a-b46d-bfe264defc14","Type":"ContainerStarted","Data":"5eb47badaaaee2e82b5b5954c92292a5671a8af6fba18795ab0e9b52440e7a70"} Oct 04 05:02:47 crc kubenswrapper[4575]: I1004 05:02:47.673021 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" event={"ID":"7a3a20d3-21b8-456a-b46d-bfe264defc14","Type":"ContainerStarted","Data":"4efa398f3e6c8b3a2529c8d53742e5cf0f1cea7f624082ed87b78de07442f069"} Oct 04 05:02:47 crc kubenswrapper[4575]: I1004 05:02:47.690669 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" podStartSLOduration=2.106191309 podStartE2EDuration="2.690651811s" podCreationTimestamp="2025-10-04 05:02:45 +0000 UTC" firstStartedPulling="2025-10-04 05:02:46.632848852 +0000 UTC m=+1717.961407666" lastFinishedPulling="2025-10-04 05:02:47.217309344 +0000 UTC m=+1718.545868168" observedRunningTime="2025-10-04 05:02:47.688698334 +0000 UTC m=+1719.017257158" watchObservedRunningTime="2025-10-04 05:02:47.690651811 +0000 UTC m=+1719.019210625" Oct 04 05:02:50 crc kubenswrapper[4575]: I1004 05:02:50.725424 4575 generic.go:334] "Generic (PLEG): container finished" podID="7a3a20d3-21b8-456a-b46d-bfe264defc14" containerID="4efa398f3e6c8b3a2529c8d53742e5cf0f1cea7f624082ed87b78de07442f069" exitCode=0 Oct 04 05:02:50 crc kubenswrapper[4575]: I1004 05:02:50.725489 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" event={"ID":"7a3a20d3-21b8-456a-b46d-bfe264defc14","Type":"ContainerDied","Data":"4efa398f3e6c8b3a2529c8d53742e5cf0f1cea7f624082ed87b78de07442f069"} Oct 04 05:02:51 crc kubenswrapper[4575]: I1004 05:02:51.309812 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:02:51 crc kubenswrapper[4575]: E1004 05:02:51.310293 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.265141 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.408990 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jswcd\" (UniqueName: \"kubernetes.io/projected/7a3a20d3-21b8-456a-b46d-bfe264defc14-kube-api-access-jswcd\") pod \"7a3a20d3-21b8-456a-b46d-bfe264defc14\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.409039 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-inventory\") pod \"7a3a20d3-21b8-456a-b46d-bfe264defc14\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.409071 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-ssh-key\") pod \"7a3a20d3-21b8-456a-b46d-bfe264defc14\" (UID: \"7a3a20d3-21b8-456a-b46d-bfe264defc14\") " Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.420387 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a3a20d3-21b8-456a-b46d-bfe264defc14-kube-api-access-jswcd" (OuterVolumeSpecName: "kube-api-access-jswcd") pod "7a3a20d3-21b8-456a-b46d-bfe264defc14" (UID: "7a3a20d3-21b8-456a-b46d-bfe264defc14"). InnerVolumeSpecName "kube-api-access-jswcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.441251 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7a3a20d3-21b8-456a-b46d-bfe264defc14" (UID: "7a3a20d3-21b8-456a-b46d-bfe264defc14"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.446647 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-inventory" (OuterVolumeSpecName: "inventory") pod "7a3a20d3-21b8-456a-b46d-bfe264defc14" (UID: "7a3a20d3-21b8-456a-b46d-bfe264defc14"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.512651 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jswcd\" (UniqueName: \"kubernetes.io/projected/7a3a20d3-21b8-456a-b46d-bfe264defc14-kube-api-access-jswcd\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.512710 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.512724 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a3a20d3-21b8-456a-b46d-bfe264defc14-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.759080 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" event={"ID":"7a3a20d3-21b8-456a-b46d-bfe264defc14","Type":"ContainerDied","Data":"5eb47badaaaee2e82b5b5954c92292a5671a8af6fba18795ab0e9b52440e7a70"} Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.759392 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eb47badaaaee2e82b5b5954c92292a5671a8af6fba18795ab0e9b52440e7a70" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.759174 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-25bjt" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.869678 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp"] Oct 04 05:02:52 crc kubenswrapper[4575]: E1004 05:02:52.870177 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3a20d3-21b8-456a-b46d-bfe264defc14" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.870199 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3a20d3-21b8-456a-b46d-bfe264defc14" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.870502 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a3a20d3-21b8-456a-b46d-bfe264defc14" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.871339 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.875008 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.875306 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.875329 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.889868 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:02:52 crc kubenswrapper[4575]: I1004 05:02:52.901446 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp"] Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.021698 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.021801 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r4vw\" (UniqueName: \"kubernetes.io/projected/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-kube-api-access-6r4vw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.021888 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.021923 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.123928 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.124001 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.124061 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.124125 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r4vw\" (UniqueName: \"kubernetes.io/projected/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-kube-api-access-6r4vw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.128295 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.143155 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.144028 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.149292 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r4vw\" (UniqueName: \"kubernetes.io/projected/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-kube-api-access-6r4vw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.188109 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.547966 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp"] Oct 04 05:02:53 crc kubenswrapper[4575]: W1004 05:02:53.554739 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c487888_8b86_4aa3_8aaf_19f4ad21f44c.slice/crio-740a7bdca6af2ff25fe3592564b2de0e06f4207e56a2f3ae839f6e3ca9150c35 WatchSource:0}: Error finding container 740a7bdca6af2ff25fe3592564b2de0e06f4207e56a2f3ae839f6e3ca9150c35: Status 404 returned error can't find the container with id 740a7bdca6af2ff25fe3592564b2de0e06f4207e56a2f3ae839f6e3ca9150c35 Oct 04 05:02:53 crc kubenswrapper[4575]: I1004 05:02:53.769446 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" event={"ID":"2c487888-8b86-4aa3-8aaf-19f4ad21f44c","Type":"ContainerStarted","Data":"740a7bdca6af2ff25fe3592564b2de0e06f4207e56a2f3ae839f6e3ca9150c35"} Oct 04 05:02:54 crc kubenswrapper[4575]: I1004 05:02:54.781384 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" event={"ID":"2c487888-8b86-4aa3-8aaf-19f4ad21f44c","Type":"ContainerStarted","Data":"79e2155a993453aba7b22ef01c5af1e30c01610b6eb372bf09175c0a572839bc"} Oct 04 05:02:54 crc kubenswrapper[4575]: I1004 05:02:54.803231 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" podStartSLOduration=1.989063354 podStartE2EDuration="2.803212416s" podCreationTimestamp="2025-10-04 05:02:52 +0000 UTC" firstStartedPulling="2025-10-04 05:02:53.558766887 +0000 UTC m=+1724.887325701" lastFinishedPulling="2025-10-04 05:02:54.372915949 +0000 UTC m=+1725.701474763" observedRunningTime="2025-10-04 05:02:54.799105256 +0000 UTC m=+1726.127664090" watchObservedRunningTime="2025-10-04 05:02:54.803212416 +0000 UTC m=+1726.131771230" Oct 04 05:03:05 crc kubenswrapper[4575]: I1004 05:03:05.311030 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:03:05 crc kubenswrapper[4575]: E1004 05:03:05.312285 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:03:20 crc kubenswrapper[4575]: I1004 05:03:20.310491 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:03:20 crc kubenswrapper[4575]: E1004 05:03:20.311420 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:03:28 crc kubenswrapper[4575]: I1004 05:03:28.241255 4575 scope.go:117] "RemoveContainer" containerID="3230dea0554883a20208ba0e92722774fde3f04a122f3182cc3cf4da62c25391" Oct 04 05:03:28 crc kubenswrapper[4575]: I1004 05:03:28.267893 4575 scope.go:117] "RemoveContainer" containerID="b47b33215697e5938212f2533d8b4e6c47cacdff986366950480273dcca9263d" Oct 04 05:03:34 crc kubenswrapper[4575]: I1004 05:03:34.310209 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:03:34 crc kubenswrapper[4575]: E1004 05:03:34.310969 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:03:45 crc kubenswrapper[4575]: I1004 05:03:45.310772 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:03:45 crc kubenswrapper[4575]: E1004 05:03:45.311456 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:03:57 crc kubenswrapper[4575]: I1004 05:03:57.310246 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:03:57 crc kubenswrapper[4575]: E1004 05:03:57.311062 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:04:08 crc kubenswrapper[4575]: I1004 05:04:08.310820 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:04:08 crc kubenswrapper[4575]: E1004 05:04:08.311652 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:04:21 crc kubenswrapper[4575]: I1004 05:04:21.309741 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:04:21 crc kubenswrapper[4575]: E1004 05:04:21.310548 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:04:34 crc kubenswrapper[4575]: I1004 05:04:34.310790 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:04:34 crc kubenswrapper[4575]: E1004 05:04:34.311650 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:04:45 crc kubenswrapper[4575]: I1004 05:04:45.310654 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:04:45 crc kubenswrapper[4575]: E1004 05:04:45.311398 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:05:00 crc kubenswrapper[4575]: I1004 05:05:00.309696 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:05:00 crc kubenswrapper[4575]: E1004 05:05:00.310363 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:05:01 crc kubenswrapper[4575]: I1004 05:05:01.057358 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-z44gg"] Oct 04 05:05:01 crc kubenswrapper[4575]: I1004 05:05:01.071181 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-z44gg"] Oct 04 05:05:01 crc kubenswrapper[4575]: I1004 05:05:01.322347 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a9115a4-74b2-49dc-90b3-d7b95bc391f4" path="/var/lib/kubelet/pods/9a9115a4-74b2-49dc-90b3-d7b95bc391f4/volumes" Oct 04 05:05:02 crc kubenswrapper[4575]: I1004 05:05:02.036962 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-g2n66"] Oct 04 05:05:02 crc kubenswrapper[4575]: I1004 05:05:02.045825 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-g2n66"] Oct 04 05:05:03 crc kubenswrapper[4575]: I1004 05:05:03.322561 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e23643b7-318b-4390-a9c3-346237015c04" path="/var/lib/kubelet/pods/e23643b7-318b-4390-a9c3-346237015c04/volumes" Oct 04 05:05:06 crc kubenswrapper[4575]: I1004 05:05:06.037157 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-5v4gf"] Oct 04 05:05:06 crc kubenswrapper[4575]: I1004 05:05:06.050308 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-5v4gf"] Oct 04 05:05:07 crc kubenswrapper[4575]: I1004 05:05:07.321536 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41d086f3-2b5e-45e5-b580-71abdd6acd71" path="/var/lib/kubelet/pods/41d086f3-2b5e-45e5-b580-71abdd6acd71/volumes" Oct 04 05:05:08 crc kubenswrapper[4575]: I1004 05:05:08.028289 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-lprm4"] Oct 04 05:05:08 crc kubenswrapper[4575]: I1004 05:05:08.036719 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-lprm4"] Oct 04 05:05:09 crc kubenswrapper[4575]: I1004 05:05:09.056466 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-72mrw"] Oct 04 05:05:09 crc kubenswrapper[4575]: I1004 05:05:09.068902 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-72mrw"] Oct 04 05:05:09 crc kubenswrapper[4575]: I1004 05:05:09.329209 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5bb6f33-61f5-4b6f-8449-2595b8593d28" path="/var/lib/kubelet/pods/c5bb6f33-61f5-4b6f-8449-2595b8593d28/volumes" Oct 04 05:05:09 crc kubenswrapper[4575]: I1004 05:05:09.333092 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eea4fe86-b04b-44d4-bc68-361b6bf34b6a" path="/var/lib/kubelet/pods/eea4fe86-b04b-44d4-bc68-361b6bf34b6a/volumes" Oct 04 05:05:10 crc kubenswrapper[4575]: I1004 05:05:10.031764 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-kjpkv"] Oct 04 05:05:10 crc kubenswrapper[4575]: I1004 05:05:10.039484 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-kjpkv"] Oct 04 05:05:11 crc kubenswrapper[4575]: I1004 05:05:11.028197 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7c2a-account-create-qvx8q"] Oct 04 05:05:11 crc kubenswrapper[4575]: I1004 05:05:11.037720 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7c2a-account-create-qvx8q"] Oct 04 05:05:11 crc kubenswrapper[4575]: I1004 05:05:11.322339 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8933fd6b-ec1a-4c42-86a2-f09d447827b6" path="/var/lib/kubelet/pods/8933fd6b-ec1a-4c42-86a2-f09d447827b6/volumes" Oct 04 05:05:11 crc kubenswrapper[4575]: I1004 05:05:11.323394 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bedd9a1e-02e4-4fac-8f24-f45234f3aa40" path="/var/lib/kubelet/pods/bedd9a1e-02e4-4fac-8f24-f45234f3aa40/volumes" Oct 04 05:05:12 crc kubenswrapper[4575]: I1004 05:05:12.028625 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-0102-account-create-p7v69"] Oct 04 05:05:12 crc kubenswrapper[4575]: I1004 05:05:12.039128 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-0102-account-create-p7v69"] Oct 04 05:05:13 crc kubenswrapper[4575]: I1004 05:05:13.327551 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34ce4329-745c-4427-8b49-e538401c3226" path="/var/lib/kubelet/pods/34ce4329-745c-4427-8b49-e538401c3226/volumes" Oct 04 05:05:15 crc kubenswrapper[4575]: I1004 05:05:15.310309 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:05:15 crc kubenswrapper[4575]: E1004 05:05:15.310837 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:05:17 crc kubenswrapper[4575]: I1004 05:05:17.029830 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-66da-account-create-6ckcq"] Oct 04 05:05:17 crc kubenswrapper[4575]: I1004 05:05:17.038727 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-66da-account-create-6ckcq"] Oct 04 05:05:17 crc kubenswrapper[4575]: I1004 05:05:17.320383 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb097de0-50a2-44a0-a1b9-3abcb12741de" path="/var/lib/kubelet/pods/bb097de0-50a2-44a0-a1b9-3abcb12741de/volumes" Oct 04 05:05:21 crc kubenswrapper[4575]: I1004 05:05:21.036747 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3a4a-account-create-knhf9"] Oct 04 05:05:21 crc kubenswrapper[4575]: I1004 05:05:21.051369 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3a4a-account-create-knhf9"] Oct 04 05:05:21 crc kubenswrapper[4575]: I1004 05:05:21.059876 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d9bf-account-create-hg27q"] Oct 04 05:05:21 crc kubenswrapper[4575]: I1004 05:05:21.069249 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d9bf-account-create-hg27q"] Oct 04 05:05:21 crc kubenswrapper[4575]: I1004 05:05:21.322294 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="274f307d-9c54-49f6-9460-2231583ff053" path="/var/lib/kubelet/pods/274f307d-9c54-49f6-9460-2231583ff053/volumes" Oct 04 05:05:21 crc kubenswrapper[4575]: I1004 05:05:21.323015 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b301bf9a-a111-4933-8f5e-51bd3d7d0c9b" path="/var/lib/kubelet/pods/b301bf9a-a111-4933-8f5e-51bd3d7d0c9b/volumes" Oct 04 05:05:27 crc kubenswrapper[4575]: I1004 05:05:27.310513 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:05:27 crc kubenswrapper[4575]: E1004 05:05:27.311425 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.341984 4575 scope.go:117] "RemoveContainer" containerID="2e207ce1c97a78cf8c52688d7420ddfaa6a5cc1fddd41f497ee1cd16f818651f" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.372027 4575 scope.go:117] "RemoveContainer" containerID="8a62fa13e09c631a28687e22ad3927fe416005701ec969d52fd561341dfc5434" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.445804 4575 scope.go:117] "RemoveContainer" containerID="c0993731dbece24a7e8c805c5361db9d7f515ca46b84e427925da9672422c0e3" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.498796 4575 scope.go:117] "RemoveContainer" containerID="ee3d7346db2a34dc02cac047164fb61bb86d8f8eca1d120759a968d5f8b9d1f5" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.559079 4575 scope.go:117] "RemoveContainer" containerID="f4bb2f72787716c30050fdbe6afcc0302e5cd9b363fd6387254b80f9896ff6bd" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.594887 4575 scope.go:117] "RemoveContainer" containerID="d3369085759623f0837e1c0a2f233a25583887366e0576693ac6a2d4465abcc0" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.618320 4575 scope.go:117] "RemoveContainer" containerID="9b8408fdb07e2e8da5bb7a92bc9bfa21a078482b939c9cf3d3ce6d013498d2c7" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.657811 4575 scope.go:117] "RemoveContainer" containerID="4c258d1f1e011ccd5f32f3add9c872c5fc1b75422008487ab35ce83605ca4137" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.680135 4575 scope.go:117] "RemoveContainer" containerID="2c7b84617bc07cf1fca4dfa5442e1ceaeb8f429e0385dc35a6fb980d816b2835" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.704612 4575 scope.go:117] "RemoveContainer" containerID="865c212f049bf39aec2d2c254afacf51d842fae1342f99c2bf42dbfc9f22dbb8" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.728537 4575 scope.go:117] "RemoveContainer" containerID="f61d6d584e189139f2e51e9ce576f8d5d56641866bfbf7a2dd52539b5a2a704a" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.753431 4575 scope.go:117] "RemoveContainer" containerID="630e8523897b9da80ee6851314cb5db2e25064a70d6c3f27acf4a8aa7dc057bb" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.776452 4575 scope.go:117] "RemoveContainer" containerID="1d36a2c8de90600d8c1f05336ae56eae60010bb2cb80c149a9b2620f21192621" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.804483 4575 scope.go:117] "RemoveContainer" containerID="c19b3329a49f313e52803203d7cf38418ae83ef19a205159a1f994464af950f4" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.826728 4575 scope.go:117] "RemoveContainer" containerID="2c5d998293c916864824f68c020bf1c52dafb1c55451382aabf4716071c249e5" Oct 04 05:05:28 crc kubenswrapper[4575]: I1004 05:05:28.852079 4575 scope.go:117] "RemoveContainer" containerID="51d26e723c7fcb9b4ac5cad529a9a6eb65907640b8b6128aa772a959028be65d" Oct 04 05:05:35 crc kubenswrapper[4575]: I1004 05:05:35.043383 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-191e-account-create-62c26"] Oct 04 05:05:35 crc kubenswrapper[4575]: I1004 05:05:35.053369 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xmg65"] Oct 04 05:05:35 crc kubenswrapper[4575]: I1004 05:05:35.065389 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-191e-account-create-62c26"] Oct 04 05:05:35 crc kubenswrapper[4575]: I1004 05:05:35.073648 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xmg65"] Oct 04 05:05:35 crc kubenswrapper[4575]: I1004 05:05:35.327577 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="918e18c5-12f4-4bf4-97c1-8bcb8441664b" path="/var/lib/kubelet/pods/918e18c5-12f4-4bf4-97c1-8bcb8441664b/volumes" Oct 04 05:05:35 crc kubenswrapper[4575]: I1004 05:05:35.329007 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0064681-6434-46c4-b46b-2c82fc5694d0" path="/var/lib/kubelet/pods/d0064681-6434-46c4-b46b-2c82fc5694d0/volumes" Oct 04 05:05:39 crc kubenswrapper[4575]: I1004 05:05:39.318367 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:05:39 crc kubenswrapper[4575]: E1004 05:05:39.319285 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:05:50 crc kubenswrapper[4575]: I1004 05:05:50.309958 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:05:50 crc kubenswrapper[4575]: E1004 05:05:50.310824 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:06:01 crc kubenswrapper[4575]: I1004 05:06:01.309455 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:06:01 crc kubenswrapper[4575]: E1004 05:06:01.310055 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:06:14 crc kubenswrapper[4575]: I1004 05:06:14.311356 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:06:14 crc kubenswrapper[4575]: I1004 05:06:14.641996 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"ecff37b59205a74250a64b395a45822895a4e3324eeb81b77ca18757cc564c58"} Oct 04 05:06:28 crc kubenswrapper[4575]: I1004 05:06:28.051985 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-gtvfg"] Oct 04 05:06:28 crc kubenswrapper[4575]: I1004 05:06:28.062351 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-gtvfg"] Oct 04 05:06:29 crc kubenswrapper[4575]: I1004 05:06:29.185931 4575 scope.go:117] "RemoveContainer" containerID="8e745803e96988e1779f47cccad8a0e0cc731d7fe71b13770fabc2425ec358c1" Oct 04 05:06:29 crc kubenswrapper[4575]: I1004 05:06:29.222702 4575 scope.go:117] "RemoveContainer" containerID="e716ad0d9bb072bfce3ab88e92d73af8a20f5e11423c4c218f2b0d79a1c43ca7" Oct 04 05:06:29 crc kubenswrapper[4575]: I1004 05:06:29.329200 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="775d96cb-5e3b-437f-b61d-7799e86f2f58" path="/var/lib/kubelet/pods/775d96cb-5e3b-437f-b61d-7799e86f2f58/volumes" Oct 04 05:06:31 crc kubenswrapper[4575]: I1004 05:06:31.816469 4575 generic.go:334] "Generic (PLEG): container finished" podID="2c487888-8b86-4aa3-8aaf-19f4ad21f44c" containerID="79e2155a993453aba7b22ef01c5af1e30c01610b6eb372bf09175c0a572839bc" exitCode=0 Oct 04 05:06:31 crc kubenswrapper[4575]: I1004 05:06:31.816581 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" event={"ID":"2c487888-8b86-4aa3-8aaf-19f4ad21f44c","Type":"ContainerDied","Data":"79e2155a993453aba7b22ef01c5af1e30c01610b6eb372bf09175c0a572839bc"} Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.261462 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.400772 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-ssh-key\") pod \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.400877 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r4vw\" (UniqueName: \"kubernetes.io/projected/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-kube-api-access-6r4vw\") pod \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.401002 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-inventory\") pod \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.401053 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-bootstrap-combined-ca-bundle\") pod \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\" (UID: \"2c487888-8b86-4aa3-8aaf-19f4ad21f44c\") " Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.407144 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2c487888-8b86-4aa3-8aaf-19f4ad21f44c" (UID: "2c487888-8b86-4aa3-8aaf-19f4ad21f44c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.407284 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-kube-api-access-6r4vw" (OuterVolumeSpecName: "kube-api-access-6r4vw") pod "2c487888-8b86-4aa3-8aaf-19f4ad21f44c" (UID: "2c487888-8b86-4aa3-8aaf-19f4ad21f44c"). InnerVolumeSpecName "kube-api-access-6r4vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.437805 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2c487888-8b86-4aa3-8aaf-19f4ad21f44c" (UID: "2c487888-8b86-4aa3-8aaf-19f4ad21f44c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.438717 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-inventory" (OuterVolumeSpecName: "inventory") pod "2c487888-8b86-4aa3-8aaf-19f4ad21f44c" (UID: "2c487888-8b86-4aa3-8aaf-19f4ad21f44c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.503385 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.503454 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r4vw\" (UniqueName: \"kubernetes.io/projected/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-kube-api-access-6r4vw\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.503466 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.503478 4575 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c487888-8b86-4aa3-8aaf-19f4ad21f44c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.838963 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" event={"ID":"2c487888-8b86-4aa3-8aaf-19f4ad21f44c","Type":"ContainerDied","Data":"740a7bdca6af2ff25fe3592564b2de0e06f4207e56a2f3ae839f6e3ca9150c35"} Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.839012 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="740a7bdca6af2ff25fe3592564b2de0e06f4207e56a2f3ae839f6e3ca9150c35" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.839078 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.933396 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl"] Oct 04 05:06:33 crc kubenswrapper[4575]: E1004 05:06:33.933862 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c487888-8b86-4aa3-8aaf-19f4ad21f44c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.933882 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c487888-8b86-4aa3-8aaf-19f4ad21f44c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.934091 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c487888-8b86-4aa3-8aaf-19f4ad21f44c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.934817 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.936739 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.950343 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.952425 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.952496 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:06:33 crc kubenswrapper[4575]: I1004 05:06:33.959467 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl"] Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.016214 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.016575 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxcc9\" (UniqueName: \"kubernetes.io/projected/a6d82a60-66ad-4718-ab96-90a5889aeb18-kube-api-access-cxcc9\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.016645 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.120555 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.120964 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxcc9\" (UniqueName: \"kubernetes.io/projected/a6d82a60-66ad-4718-ab96-90a5889aeb18-kube-api-access-cxcc9\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.121053 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.141197 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.147754 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.159740 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxcc9\" (UniqueName: \"kubernetes.io/projected/a6d82a60-66ad-4718-ab96-90a5889aeb18-kube-api-access-cxcc9\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.254147 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.836648 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl"] Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.846980 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:06:34 crc kubenswrapper[4575]: I1004 05:06:34.855248 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" event={"ID":"a6d82a60-66ad-4718-ab96-90a5889aeb18","Type":"ContainerStarted","Data":"e9e5fe5882d0f99f7597188b0546353f14d3bf6f3de6893e905074e307b5d621"} Oct 04 05:06:35 crc kubenswrapper[4575]: I1004 05:06:35.866170 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" event={"ID":"a6d82a60-66ad-4718-ab96-90a5889aeb18","Type":"ContainerStarted","Data":"bd24b2d0b7041d47270f5bc4f3a7277ece5aa5c620f1e77335f04fe8d8485d47"} Oct 04 05:06:35 crc kubenswrapper[4575]: I1004 05:06:35.890212 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" podStartSLOduration=2.425896135 podStartE2EDuration="2.890187802s" podCreationTimestamp="2025-10-04 05:06:33 +0000 UTC" firstStartedPulling="2025-10-04 05:06:34.846764711 +0000 UTC m=+1946.175323525" lastFinishedPulling="2025-10-04 05:06:35.311056378 +0000 UTC m=+1946.639615192" observedRunningTime="2025-10-04 05:06:35.887566956 +0000 UTC m=+1947.216125790" watchObservedRunningTime="2025-10-04 05:06:35.890187802 +0000 UTC m=+1947.218746626" Oct 04 05:06:40 crc kubenswrapper[4575]: I1004 05:06:40.054056 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-459j5"] Oct 04 05:06:40 crc kubenswrapper[4575]: I1004 05:06:40.068500 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-459j5"] Oct 04 05:06:41 crc kubenswrapper[4575]: I1004 05:06:41.031802 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-99bks"] Oct 04 05:06:41 crc kubenswrapper[4575]: I1004 05:06:41.041999 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-6qj89"] Oct 04 05:06:41 crc kubenswrapper[4575]: I1004 05:06:41.052207 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-99bks"] Oct 04 05:06:41 crc kubenswrapper[4575]: I1004 05:06:41.061787 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-6qj89"] Oct 04 05:06:41 crc kubenswrapper[4575]: I1004 05:06:41.323159 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62f6c233-faeb-47e1-8103-4902eaea8d4a" path="/var/lib/kubelet/pods/62f6c233-faeb-47e1-8103-4902eaea8d4a/volumes" Oct 04 05:06:41 crc kubenswrapper[4575]: I1004 05:06:41.324084 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca934f38-b0b0-4d26-bf2e-63f983c81851" path="/var/lib/kubelet/pods/ca934f38-b0b0-4d26-bf2e-63f983c81851/volumes" Oct 04 05:06:41 crc kubenswrapper[4575]: I1004 05:06:41.324852 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc2d4d08-1560-4aa3-8a6e-5350abdd3b92" path="/var/lib/kubelet/pods/fc2d4d08-1560-4aa3-8a6e-5350abdd3b92/volumes" Oct 04 05:07:03 crc kubenswrapper[4575]: I1004 05:07:03.047285 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-kg5nx"] Oct 04 05:07:03 crc kubenswrapper[4575]: I1004 05:07:03.055384 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-kg5nx"] Oct 04 05:07:03 crc kubenswrapper[4575]: I1004 05:07:03.320028 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef74f3cc-9958-4de3-a914-12e5453e169d" path="/var/lib/kubelet/pods/ef74f3cc-9958-4de3-a914-12e5453e169d/volumes" Oct 04 05:07:07 crc kubenswrapper[4575]: I1004 05:07:07.077490 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-7kr5n"] Oct 04 05:07:07 crc kubenswrapper[4575]: I1004 05:07:07.096016 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-7kr5n"] Oct 04 05:07:07 crc kubenswrapper[4575]: I1004 05:07:07.322618 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d089c375-6e05-4d74-88e1-d00c8cbddb1f" path="/var/lib/kubelet/pods/d089c375-6e05-4d74-88e1-d00c8cbddb1f/volumes" Oct 04 05:07:29 crc kubenswrapper[4575]: I1004 05:07:29.335132 4575 scope.go:117] "RemoveContainer" containerID="a45fe580eb50011e8314d6a4e9d7459efe89b3417282abc0f07b1360280af6d2" Oct 04 05:07:29 crc kubenswrapper[4575]: I1004 05:07:29.385668 4575 scope.go:117] "RemoveContainer" containerID="dd502cb233d95c2c516c52699cbfd531bb3844e95479f6d08ba7fc1bc7d0794a" Oct 04 05:07:29 crc kubenswrapper[4575]: I1004 05:07:29.439317 4575 scope.go:117] "RemoveContainer" containerID="b88c3e2d3ac777ec98e5b700d75c1cfaf4895dd17819590c14bf1ffc0030ec62" Oct 04 05:07:29 crc kubenswrapper[4575]: I1004 05:07:29.510316 4575 scope.go:117] "RemoveContainer" containerID="4b93a345b80763c54704d57faa06830fee0618b5996fd0dd384828d49fd5a1c3" Oct 04 05:07:29 crc kubenswrapper[4575]: I1004 05:07:29.550505 4575 scope.go:117] "RemoveContainer" containerID="3bd213d96e5e86bf506ad4db79bad1f7fde3805d418daa4d347cfcd31d0ed19e" Oct 04 05:07:29 crc kubenswrapper[4575]: I1004 05:07:29.609139 4575 scope.go:117] "RemoveContainer" containerID="2dab7ad46d64d8b960c4f7b4044c68c2595dbab9cc061d49f6dd8f075dc41c0c" Oct 04 05:08:14 crc kubenswrapper[4575]: I1004 05:08:14.053745 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-j94zx"] Oct 04 05:08:14 crc kubenswrapper[4575]: I1004 05:08:14.065283 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-fh56z"] Oct 04 05:08:14 crc kubenswrapper[4575]: I1004 05:08:14.074612 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-bcghq"] Oct 04 05:08:14 crc kubenswrapper[4575]: I1004 05:08:14.087544 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-bcghq"] Oct 04 05:08:14 crc kubenswrapper[4575]: I1004 05:08:14.096519 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-fh56z"] Oct 04 05:08:14 crc kubenswrapper[4575]: I1004 05:08:14.107191 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-j94zx"] Oct 04 05:08:15 crc kubenswrapper[4575]: I1004 05:08:15.321276 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c3e070b-a4ff-4220-a4f9-f9750733357f" path="/var/lib/kubelet/pods/6c3e070b-a4ff-4220-a4f9-f9750733357f/volumes" Oct 04 05:08:15 crc kubenswrapper[4575]: I1004 05:08:15.322407 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99b2d0ef-02b1-4a22-b4c0-475d94076370" path="/var/lib/kubelet/pods/99b2d0ef-02b1-4a22-b4c0-475d94076370/volumes" Oct 04 05:08:15 crc kubenswrapper[4575]: I1004 05:08:15.323637 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b" path="/var/lib/kubelet/pods/bf2ed6e4-59f3-4c97-ab79-abeb0dd1e37b/volumes" Oct 04 05:08:22 crc kubenswrapper[4575]: I1004 05:08:22.885016 4575 generic.go:334] "Generic (PLEG): container finished" podID="a6d82a60-66ad-4718-ab96-90a5889aeb18" containerID="bd24b2d0b7041d47270f5bc4f3a7277ece5aa5c620f1e77335f04fe8d8485d47" exitCode=0 Oct 04 05:08:22 crc kubenswrapper[4575]: I1004 05:08:22.885466 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" event={"ID":"a6d82a60-66ad-4718-ab96-90a5889aeb18","Type":"ContainerDied","Data":"bd24b2d0b7041d47270f5bc4f3a7277ece5aa5c620f1e77335f04fe8d8485d47"} Oct 04 05:08:23 crc kubenswrapper[4575]: I1004 05:08:23.041951 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-aff9-account-create-42nv4"] Oct 04 05:08:23 crc kubenswrapper[4575]: I1004 05:08:23.052786 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-aff9-account-create-42nv4"] Oct 04 05:08:23 crc kubenswrapper[4575]: I1004 05:08:23.322309 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de414a31-468d-4a87-9aae-26e3b05702da" path="/var/lib/kubelet/pods/de414a31-468d-4a87-9aae-26e3b05702da/volumes" Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.032767 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f3a6-account-create-7kb28"] Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.043150 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f3a6-account-create-7kb28"] Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.050552 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-3e15-account-create-jcnqg"] Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.057566 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-3e15-account-create-jcnqg"] Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.333262 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.469978 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-inventory\") pod \"a6d82a60-66ad-4718-ab96-90a5889aeb18\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.470077 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-ssh-key\") pod \"a6d82a60-66ad-4718-ab96-90a5889aeb18\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.470186 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxcc9\" (UniqueName: \"kubernetes.io/projected/a6d82a60-66ad-4718-ab96-90a5889aeb18-kube-api-access-cxcc9\") pod \"a6d82a60-66ad-4718-ab96-90a5889aeb18\" (UID: \"a6d82a60-66ad-4718-ab96-90a5889aeb18\") " Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.479538 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6d82a60-66ad-4718-ab96-90a5889aeb18-kube-api-access-cxcc9" (OuterVolumeSpecName: "kube-api-access-cxcc9") pod "a6d82a60-66ad-4718-ab96-90a5889aeb18" (UID: "a6d82a60-66ad-4718-ab96-90a5889aeb18"). InnerVolumeSpecName "kube-api-access-cxcc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.499230 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a6d82a60-66ad-4718-ab96-90a5889aeb18" (UID: "a6d82a60-66ad-4718-ab96-90a5889aeb18"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.506101 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-inventory" (OuterVolumeSpecName: "inventory") pod "a6d82a60-66ad-4718-ab96-90a5889aeb18" (UID: "a6d82a60-66ad-4718-ab96-90a5889aeb18"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.575000 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.575046 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6d82a60-66ad-4718-ab96-90a5889aeb18-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.575059 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxcc9\" (UniqueName: \"kubernetes.io/projected/a6d82a60-66ad-4718-ab96-90a5889aeb18-kube-api-access-cxcc9\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.914707 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" event={"ID":"a6d82a60-66ad-4718-ab96-90a5889aeb18","Type":"ContainerDied","Data":"e9e5fe5882d0f99f7597188b0546353f14d3bf6f3de6893e905074e307b5d621"} Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.914795 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e5fe5882d0f99f7597188b0546353f14d3bf6f3de6893e905074e307b5d621" Oct 04 05:08:24 crc kubenswrapper[4575]: I1004 05:08:24.915236 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.014532 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd"] Oct 04 05:08:25 crc kubenswrapper[4575]: E1004 05:08:25.014948 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d82a60-66ad-4718-ab96-90a5889aeb18" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.014965 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d82a60-66ad-4718-ab96-90a5889aeb18" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.015174 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d82a60-66ad-4718-ab96-90a5889aeb18" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.015826 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.019481 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.020039 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.020205 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.023277 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.030473 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd"] Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.084858 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n48x\" (UniqueName: \"kubernetes.io/projected/583857d7-9e34-4be2-9376-b7261f940d9e-kube-api-access-9n48x\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.084992 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.085063 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.186352 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n48x\" (UniqueName: \"kubernetes.io/projected/583857d7-9e34-4be2-9376-b7261f940d9e-kube-api-access-9n48x\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.186484 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.186534 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.191193 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.201233 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.211972 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n48x\" (UniqueName: \"kubernetes.io/projected/583857d7-9e34-4be2-9376-b7261f940d9e-kube-api-access-9n48x\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.323681 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210722f0-a287-4f59-9fb2-8653ec84310e" path="/var/lib/kubelet/pods/210722f0-a287-4f59-9fb2-8653ec84310e/volumes" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.324721 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a72b81d1-ed82-47ea-911f-03c05e5f3d32" path="/var/lib/kubelet/pods/a72b81d1-ed82-47ea-911f-03c05e5f3d32/volumes" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.339889 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.885769 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd"] Oct 04 05:08:25 crc kubenswrapper[4575]: I1004 05:08:25.925065 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" event={"ID":"583857d7-9e34-4be2-9376-b7261f940d9e","Type":"ContainerStarted","Data":"0cbeafbaff8662a826ba914631cd6a4c7f2fa2bf147b7327b3480ff91126ce23"} Oct 04 05:08:26 crc kubenswrapper[4575]: I1004 05:08:26.939525 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" event={"ID":"583857d7-9e34-4be2-9376-b7261f940d9e","Type":"ContainerStarted","Data":"a9ae60066ecc959ea46c401a2e27dfe48fad47bfadb1a83e12e847497909402e"} Oct 04 05:08:27 crc kubenswrapper[4575]: I1004 05:08:27.973146 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" podStartSLOduration=3.215562357 podStartE2EDuration="3.973125419s" podCreationTimestamp="2025-10-04 05:08:24 +0000 UTC" firstStartedPulling="2025-10-04 05:08:25.895038722 +0000 UTC m=+2057.223597536" lastFinishedPulling="2025-10-04 05:08:26.652601784 +0000 UTC m=+2057.981160598" observedRunningTime="2025-10-04 05:08:27.964052996 +0000 UTC m=+2059.292611810" watchObservedRunningTime="2025-10-04 05:08:27.973125419 +0000 UTC m=+2059.301684233" Oct 04 05:08:29 crc kubenswrapper[4575]: I1004 05:08:29.803281 4575 scope.go:117] "RemoveContainer" containerID="d849fe92b5f9de62575504a96a7948d0429ca8638ee4d1fbc3925e4cc0ea9a6d" Oct 04 05:08:29 crc kubenswrapper[4575]: I1004 05:08:29.828985 4575 scope.go:117] "RemoveContainer" containerID="1e273f1c6f0ed767b53473126090e37faf54ab31ec2dd34c40461d322eb6ff2e" Oct 04 05:08:29 crc kubenswrapper[4575]: I1004 05:08:29.875930 4575 scope.go:117] "RemoveContainer" containerID="9dbb8fed00babf10448b6214c199fbd0bc1d6dfbd9cad3a5741b30d153dddc0b" Oct 04 05:08:29 crc kubenswrapper[4575]: I1004 05:08:29.918161 4575 scope.go:117] "RemoveContainer" containerID="dd9ec1e13da84c8c60dcc935fca3a6e3979137ceee1d35caf019987f29048439" Oct 04 05:08:29 crc kubenswrapper[4575]: I1004 05:08:29.961617 4575 scope.go:117] "RemoveContainer" containerID="1f374dc69a5553d5309029899bf15a1a16162ddb4f17171c2e9346d729cf8ffa" Oct 04 05:08:30 crc kubenswrapper[4575]: I1004 05:08:30.006393 4575 scope.go:117] "RemoveContainer" containerID="71ab791e8d7de5acca7be2417dd76e530b128e98457332747480d96154161d1d" Oct 04 05:08:38 crc kubenswrapper[4575]: I1004 05:08:38.446049 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:08:38 crc kubenswrapper[4575]: I1004 05:08:38.446729 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:08:50 crc kubenswrapper[4575]: I1004 05:08:50.043777 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2s2b6"] Oct 04 05:08:50 crc kubenswrapper[4575]: I1004 05:08:50.053189 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2s2b6"] Oct 04 05:08:51 crc kubenswrapper[4575]: I1004 05:08:51.320960 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8d694b1-ab26-49ff-97c6-4a26ef3a0142" path="/var/lib/kubelet/pods/f8d694b1-ab26-49ff-97c6-4a26ef3a0142/volumes" Oct 04 05:09:08 crc kubenswrapper[4575]: I1004 05:09:08.446633 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:09:08 crc kubenswrapper[4575]: I1004 05:09:08.447175 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:09:23 crc kubenswrapper[4575]: I1004 05:09:23.047385 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-pldbk"] Oct 04 05:09:23 crc kubenswrapper[4575]: I1004 05:09:23.058404 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-pldbk"] Oct 04 05:09:23 crc kubenswrapper[4575]: I1004 05:09:23.320575 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8" path="/var/lib/kubelet/pods/851e6d89-cb88-4f67-b3ce-0ad1a3b0e7b8/volumes" Oct 04 05:09:25 crc kubenswrapper[4575]: I1004 05:09:25.026126 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-znqtq"] Oct 04 05:09:25 crc kubenswrapper[4575]: I1004 05:09:25.034869 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-znqtq"] Oct 04 05:09:25 crc kubenswrapper[4575]: I1004 05:09:25.321123 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74" path="/var/lib/kubelet/pods/6d84a9d7-e9be-4cb9-a6bd-9df57dbb2d74/volumes" Oct 04 05:09:30 crc kubenswrapper[4575]: I1004 05:09:30.188903 4575 scope.go:117] "RemoveContainer" containerID="5ba8d76ad793e81069c3f3e732151ec619db7ced0cd6d6118b6effe957a41c5e" Oct 04 05:09:30 crc kubenswrapper[4575]: I1004 05:09:30.241224 4575 scope.go:117] "RemoveContainer" containerID="e113320cd2896101d1b308ab535199ca02eabed2946ee283bc62645c482dcc72" Oct 04 05:09:30 crc kubenswrapper[4575]: I1004 05:09:30.295744 4575 scope.go:117] "RemoveContainer" containerID="92ff953ac58e7b34212b6fb3f6650f4156d0273132432a20d45b37fd75280541" Oct 04 05:09:38 crc kubenswrapper[4575]: I1004 05:09:38.446041 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:09:38 crc kubenswrapper[4575]: I1004 05:09:38.447055 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:09:38 crc kubenswrapper[4575]: I1004 05:09:38.447134 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:09:38 crc kubenswrapper[4575]: I1004 05:09:38.448209 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ecff37b59205a74250a64b395a45822895a4e3324eeb81b77ca18757cc564c58"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:09:38 crc kubenswrapper[4575]: I1004 05:09:38.448319 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://ecff37b59205a74250a64b395a45822895a4e3324eeb81b77ca18757cc564c58" gracePeriod=600 Oct 04 05:09:39 crc kubenswrapper[4575]: I1004 05:09:39.573422 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="ecff37b59205a74250a64b395a45822895a4e3324eeb81b77ca18757cc564c58" exitCode=0 Oct 04 05:09:39 crc kubenswrapper[4575]: I1004 05:09:39.573469 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"ecff37b59205a74250a64b395a45822895a4e3324eeb81b77ca18757cc564c58"} Oct 04 05:09:39 crc kubenswrapper[4575]: I1004 05:09:39.574043 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e"} Oct 04 05:09:39 crc kubenswrapper[4575]: I1004 05:09:39.574069 4575 scope.go:117] "RemoveContainer" containerID="f4bce908eeecc89e8ceb3642283a3a0d19d7419911a43622304552dfbf50bc22" Oct 04 05:09:40 crc kubenswrapper[4575]: I1004 05:09:40.587006 4575 generic.go:334] "Generic (PLEG): container finished" podID="583857d7-9e34-4be2-9376-b7261f940d9e" containerID="a9ae60066ecc959ea46c401a2e27dfe48fad47bfadb1a83e12e847497909402e" exitCode=0 Oct 04 05:09:40 crc kubenswrapper[4575]: I1004 05:09:40.587089 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" event={"ID":"583857d7-9e34-4be2-9376-b7261f940d9e","Type":"ContainerDied","Data":"a9ae60066ecc959ea46c401a2e27dfe48fad47bfadb1a83e12e847497909402e"} Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.043794 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.078402 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-inventory\") pod \"583857d7-9e34-4be2-9376-b7261f940d9e\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.079059 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n48x\" (UniqueName: \"kubernetes.io/projected/583857d7-9e34-4be2-9376-b7261f940d9e-kube-api-access-9n48x\") pod \"583857d7-9e34-4be2-9376-b7261f940d9e\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.079160 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-ssh-key\") pod \"583857d7-9e34-4be2-9376-b7261f940d9e\" (UID: \"583857d7-9e34-4be2-9376-b7261f940d9e\") " Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.086362 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583857d7-9e34-4be2-9376-b7261f940d9e-kube-api-access-9n48x" (OuterVolumeSpecName: "kube-api-access-9n48x") pod "583857d7-9e34-4be2-9376-b7261f940d9e" (UID: "583857d7-9e34-4be2-9376-b7261f940d9e"). InnerVolumeSpecName "kube-api-access-9n48x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.115688 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "583857d7-9e34-4be2-9376-b7261f940d9e" (UID: "583857d7-9e34-4be2-9376-b7261f940d9e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.119509 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-inventory" (OuterVolumeSpecName: "inventory") pod "583857d7-9e34-4be2-9376-b7261f940d9e" (UID: "583857d7-9e34-4be2-9376-b7261f940d9e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.181852 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.181898 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n48x\" (UniqueName: \"kubernetes.io/projected/583857d7-9e34-4be2-9376-b7261f940d9e-kube-api-access-9n48x\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.181914 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/583857d7-9e34-4be2-9376-b7261f940d9e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.196271 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8tsw6"] Oct 04 05:09:42 crc kubenswrapper[4575]: E1004 05:09:42.196933 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583857d7-9e34-4be2-9376-b7261f940d9e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.196961 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="583857d7-9e34-4be2-9376-b7261f940d9e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.197386 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="583857d7-9e34-4be2-9376-b7261f940d9e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.198848 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.262775 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8tsw6"] Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.282912 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-utilities\") pod \"certified-operators-8tsw6\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.283027 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grzjd\" (UniqueName: \"kubernetes.io/projected/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-kube-api-access-grzjd\") pod \"certified-operators-8tsw6\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.283074 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-catalog-content\") pod \"certified-operators-8tsw6\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.385177 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grzjd\" (UniqueName: \"kubernetes.io/projected/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-kube-api-access-grzjd\") pod \"certified-operators-8tsw6\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.385295 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-catalog-content\") pod \"certified-operators-8tsw6\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.385474 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-utilities\") pod \"certified-operators-8tsw6\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.385965 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-utilities\") pod \"certified-operators-8tsw6\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.386095 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-catalog-content\") pod \"certified-operators-8tsw6\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.410547 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grzjd\" (UniqueName: \"kubernetes.io/projected/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-kube-api-access-grzjd\") pod \"certified-operators-8tsw6\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.565667 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.616026 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" event={"ID":"583857d7-9e34-4be2-9376-b7261f940d9e","Type":"ContainerDied","Data":"0cbeafbaff8662a826ba914631cd6a4c7f2fa2bf147b7327b3480ff91126ce23"} Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.616177 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.616060 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cbeafbaff8662a826ba914631cd6a4c7f2fa2bf147b7327b3480ff91126ce23" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.753080 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp"] Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.755912 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.764021 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.764055 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.764276 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.764303 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.772976 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp"] Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.791646 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqr25\" (UniqueName: \"kubernetes.io/projected/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-kube-api-access-jqr25\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-97tnp\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.791697 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-97tnp\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.791830 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-97tnp\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.893257 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-97tnp\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.893421 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqr25\" (UniqueName: \"kubernetes.io/projected/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-kube-api-access-jqr25\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-97tnp\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.893462 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-97tnp\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.900278 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-97tnp\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.904758 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-97tnp\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:42 crc kubenswrapper[4575]: I1004 05:09:42.915956 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqr25\" (UniqueName: \"kubernetes.io/projected/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-kube-api-access-jqr25\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-97tnp\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:43 crc kubenswrapper[4575]: I1004 05:09:43.094628 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:43 crc kubenswrapper[4575]: I1004 05:09:43.175719 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8tsw6"] Oct 04 05:09:43 crc kubenswrapper[4575]: I1004 05:09:43.626132 4575 generic.go:334] "Generic (PLEG): container finished" podID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerID="bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27" exitCode=0 Oct 04 05:09:43 crc kubenswrapper[4575]: I1004 05:09:43.626184 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tsw6" event={"ID":"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c","Type":"ContainerDied","Data":"bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27"} Oct 04 05:09:43 crc kubenswrapper[4575]: I1004 05:09:43.626713 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tsw6" event={"ID":"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c","Type":"ContainerStarted","Data":"f4b79195c5badf7ab21387c16567cf309000a34d577a58170a1c25418d79a7c2"} Oct 04 05:09:43 crc kubenswrapper[4575]: I1004 05:09:43.652056 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp"] Oct 04 05:09:44 crc kubenswrapper[4575]: I1004 05:09:44.638905 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" event={"ID":"89ee882d-998a-4844-a65a-a1a7a1bf9bb7","Type":"ContainerStarted","Data":"881b085afedc2d0a90cc9f56f77156535f2893eb26c50d650a5fab30a5372a7f"} Oct 04 05:09:44 crc kubenswrapper[4575]: I1004 05:09:44.639529 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" event={"ID":"89ee882d-998a-4844-a65a-a1a7a1bf9bb7","Type":"ContainerStarted","Data":"4a4d9bb5984f93747ed9a03683b6bd6e31c65061ef0ef8b252beae4dc976cab5"} Oct 04 05:09:44 crc kubenswrapper[4575]: I1004 05:09:44.641488 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tsw6" event={"ID":"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c","Type":"ContainerStarted","Data":"73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123"} Oct 04 05:09:44 crc kubenswrapper[4575]: I1004 05:09:44.654736 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" podStartSLOduration=2.211313963 podStartE2EDuration="2.65471622s" podCreationTimestamp="2025-10-04 05:09:42 +0000 UTC" firstStartedPulling="2025-10-04 05:09:43.66226783 +0000 UTC m=+2134.990826644" lastFinishedPulling="2025-10-04 05:09:44.105670087 +0000 UTC m=+2135.434228901" observedRunningTime="2025-10-04 05:09:44.653302789 +0000 UTC m=+2135.981861613" watchObservedRunningTime="2025-10-04 05:09:44.65471622 +0000 UTC m=+2135.983275044" Oct 04 05:09:46 crc kubenswrapper[4575]: I1004 05:09:46.660516 4575 generic.go:334] "Generic (PLEG): container finished" podID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerID="73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123" exitCode=0 Oct 04 05:09:46 crc kubenswrapper[4575]: I1004 05:09:46.660658 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tsw6" event={"ID":"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c","Type":"ContainerDied","Data":"73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123"} Oct 04 05:09:47 crc kubenswrapper[4575]: I1004 05:09:47.672239 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tsw6" event={"ID":"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c","Type":"ContainerStarted","Data":"c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa"} Oct 04 05:09:47 crc kubenswrapper[4575]: I1004 05:09:47.696651 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8tsw6" podStartSLOduration=2.2873773 podStartE2EDuration="5.696628671s" podCreationTimestamp="2025-10-04 05:09:42 +0000 UTC" firstStartedPulling="2025-10-04 05:09:43.628307865 +0000 UTC m=+2134.956866679" lastFinishedPulling="2025-10-04 05:09:47.037559236 +0000 UTC m=+2138.366118050" observedRunningTime="2025-10-04 05:09:47.688783584 +0000 UTC m=+2139.017342418" watchObservedRunningTime="2025-10-04 05:09:47.696628671 +0000 UTC m=+2139.025187495" Oct 04 05:09:49 crc kubenswrapper[4575]: I1004 05:09:49.687498 4575 generic.go:334] "Generic (PLEG): container finished" podID="89ee882d-998a-4844-a65a-a1a7a1bf9bb7" containerID="881b085afedc2d0a90cc9f56f77156535f2893eb26c50d650a5fab30a5372a7f" exitCode=0 Oct 04 05:09:49 crc kubenswrapper[4575]: I1004 05:09:49.687619 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" event={"ID":"89ee882d-998a-4844-a65a-a1a7a1bf9bb7","Type":"ContainerDied","Data":"881b085afedc2d0a90cc9f56f77156535f2893eb26c50d650a5fab30a5372a7f"} Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.109321 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.257332 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-ssh-key\") pod \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.257431 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-inventory\") pod \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.257508 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqr25\" (UniqueName: \"kubernetes.io/projected/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-kube-api-access-jqr25\") pod \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\" (UID: \"89ee882d-998a-4844-a65a-a1a7a1bf9bb7\") " Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.263050 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-kube-api-access-jqr25" (OuterVolumeSpecName: "kube-api-access-jqr25") pod "89ee882d-998a-4844-a65a-a1a7a1bf9bb7" (UID: "89ee882d-998a-4844-a65a-a1a7a1bf9bb7"). InnerVolumeSpecName "kube-api-access-jqr25". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.288015 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-inventory" (OuterVolumeSpecName: "inventory") pod "89ee882d-998a-4844-a65a-a1a7a1bf9bb7" (UID: "89ee882d-998a-4844-a65a-a1a7a1bf9bb7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.305037 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "89ee882d-998a-4844-a65a-a1a7a1bf9bb7" (UID: "89ee882d-998a-4844-a65a-a1a7a1bf9bb7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.360487 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqr25\" (UniqueName: \"kubernetes.io/projected/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-kube-api-access-jqr25\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.360801 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.360952 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89ee882d-998a-4844-a65a-a1a7a1bf9bb7-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.704542 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" event={"ID":"89ee882d-998a-4844-a65a-a1a7a1bf9bb7","Type":"ContainerDied","Data":"4a4d9bb5984f93747ed9a03683b6bd6e31c65061ef0ef8b252beae4dc976cab5"} Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.704608 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a4d9bb5984f93747ed9a03683b6bd6e31c65061ef0ef8b252beae4dc976cab5" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.704716 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-97tnp" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.798632 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr"] Oct 04 05:09:51 crc kubenswrapper[4575]: E1004 05:09:51.799057 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89ee882d-998a-4844-a65a-a1a7a1bf9bb7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.799080 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="89ee882d-998a-4844-a65a-a1a7a1bf9bb7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.799406 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="89ee882d-998a-4844-a65a-a1a7a1bf9bb7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.800067 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.802184 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.803171 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.803309 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.803427 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.805773 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr"] Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.971159 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6gfmr\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.971209 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6gfmr\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:51 crc kubenswrapper[4575]: I1004 05:09:51.971608 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvzcg\" (UniqueName: \"kubernetes.io/projected/5734782c-1920-445b-9225-f3172b8f28c8-kube-api-access-rvzcg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6gfmr\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.074162 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6gfmr\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.074688 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6gfmr\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.075092 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvzcg\" (UniqueName: \"kubernetes.io/projected/5734782c-1920-445b-9225-f3172b8f28c8-kube-api-access-rvzcg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6gfmr\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.086871 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6gfmr\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.086994 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6gfmr\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.097047 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvzcg\" (UniqueName: \"kubernetes.io/projected/5734782c-1920-445b-9225-f3172b8f28c8-kube-api-access-rvzcg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6gfmr\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.123440 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.565947 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.566334 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:09:52 crc kubenswrapper[4575]: I1004 05:09:52.760021 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr"] Oct 04 05:09:53 crc kubenswrapper[4575]: I1004 05:09:53.629675 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8tsw6" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerName="registry-server" probeResult="failure" output=< Oct 04 05:09:53 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:09:53 crc kubenswrapper[4575]: > Oct 04 05:09:53 crc kubenswrapper[4575]: I1004 05:09:53.724523 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" event={"ID":"5734782c-1920-445b-9225-f3172b8f28c8","Type":"ContainerStarted","Data":"758d40648a3c98d2f16d6f77edcec7ef8da26a439914f24a1674cf2d689e5255"} Oct 04 05:09:53 crc kubenswrapper[4575]: I1004 05:09:53.724899 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" event={"ID":"5734782c-1920-445b-9225-f3172b8f28c8","Type":"ContainerStarted","Data":"2868a8bf0d04bff69c7ebd1042a90111230a5c1e6f8994434ecb1c406334cdba"} Oct 04 05:09:53 crc kubenswrapper[4575]: I1004 05:09:53.742567 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" podStartSLOduration=2.178384424 podStartE2EDuration="2.742546314s" podCreationTimestamp="2025-10-04 05:09:51 +0000 UTC" firstStartedPulling="2025-10-04 05:09:52.770721384 +0000 UTC m=+2144.099280198" lastFinishedPulling="2025-10-04 05:09:53.334883274 +0000 UTC m=+2144.663442088" observedRunningTime="2025-10-04 05:09:53.742432341 +0000 UTC m=+2145.070991155" watchObservedRunningTime="2025-10-04 05:09:53.742546314 +0000 UTC m=+2145.071105128" Oct 04 05:10:02 crc kubenswrapper[4575]: I1004 05:10:02.614526 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:10:02 crc kubenswrapper[4575]: I1004 05:10:02.667966 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:10:02 crc kubenswrapper[4575]: I1004 05:10:02.848180 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8tsw6"] Oct 04 05:10:03 crc kubenswrapper[4575]: I1004 05:10:03.816741 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8tsw6" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerName="registry-server" containerID="cri-o://c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa" gracePeriod=2 Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.254047 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.258090 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-utilities\") pod \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.258138 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-catalog-content\") pod \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.258167 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grzjd\" (UniqueName: \"kubernetes.io/projected/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-kube-api-access-grzjd\") pod \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\" (UID: \"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c\") " Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.259026 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-utilities" (OuterVolumeSpecName: "utilities") pod "f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" (UID: "f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.274063 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-kube-api-access-grzjd" (OuterVolumeSpecName: "kube-api-access-grzjd") pod "f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" (UID: "f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c"). InnerVolumeSpecName "kube-api-access-grzjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.333411 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" (UID: "f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.360678 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.360706 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.360716 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grzjd\" (UniqueName: \"kubernetes.io/projected/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c-kube-api-access-grzjd\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.825779 4575 generic.go:334] "Generic (PLEG): container finished" podID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerID="c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa" exitCode=0 Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.825819 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tsw6" event={"ID":"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c","Type":"ContainerDied","Data":"c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa"} Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.826110 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8tsw6" event={"ID":"f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c","Type":"ContainerDied","Data":"f4b79195c5badf7ab21387c16567cf309000a34d577a58170a1c25418d79a7c2"} Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.826135 4575 scope.go:117] "RemoveContainer" containerID="c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.825877 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8tsw6" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.854686 4575 scope.go:117] "RemoveContainer" containerID="73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.861788 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8tsw6"] Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.880894 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8tsw6"] Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.886518 4575 scope.go:117] "RemoveContainer" containerID="bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.926389 4575 scope.go:117] "RemoveContainer" containerID="c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa" Oct 04 05:10:04 crc kubenswrapper[4575]: E1004 05:10:04.926949 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa\": container with ID starting with c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa not found: ID does not exist" containerID="c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.927054 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa"} err="failed to get container status \"c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa\": rpc error: code = NotFound desc = could not find container \"c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa\": container with ID starting with c1e09be282c5aedea71ccd690a1f3b11228cb7bad656be461cb1233566c2acfa not found: ID does not exist" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.927134 4575 scope.go:117] "RemoveContainer" containerID="73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123" Oct 04 05:10:04 crc kubenswrapper[4575]: E1004 05:10:04.927546 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123\": container with ID starting with 73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123 not found: ID does not exist" containerID="73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.927688 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123"} err="failed to get container status \"73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123\": rpc error: code = NotFound desc = could not find container \"73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123\": container with ID starting with 73345a0219d510d95f9cb7c309242e47e3125381766d7ef528249568480ad123 not found: ID does not exist" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.927783 4575 scope.go:117] "RemoveContainer" containerID="bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27" Oct 04 05:10:04 crc kubenswrapper[4575]: E1004 05:10:04.929343 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27\": container with ID starting with bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27 not found: ID does not exist" containerID="bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27" Oct 04 05:10:04 crc kubenswrapper[4575]: I1004 05:10:04.929442 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27"} err="failed to get container status \"bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27\": rpc error: code = NotFound desc = could not find container \"bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27\": container with ID starting with bef5fe1fd4824db4ebb30308764acd55480ff68036cc527ebf170e59922a0e27 not found: ID does not exist" Oct 04 05:10:05 crc kubenswrapper[4575]: I1004 05:10:05.321969 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" path="/var/lib/kubelet/pods/f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c/volumes" Oct 04 05:10:08 crc kubenswrapper[4575]: I1004 05:10:08.045937 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-k4drx"] Oct 04 05:10:08 crc kubenswrapper[4575]: I1004 05:10:08.054573 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-k4drx"] Oct 04 05:10:09 crc kubenswrapper[4575]: I1004 05:10:09.327938 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d863756-7f8d-4e1c-ad16-b44be09678c1" path="/var/lib/kubelet/pods/7d863756-7f8d-4e1c-ad16-b44be09678c1/volumes" Oct 04 05:10:19 crc kubenswrapper[4575]: I1004 05:10:19.981984 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7f2jb"] Oct 04 05:10:19 crc kubenswrapper[4575]: E1004 05:10:19.983352 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerName="extract-utilities" Oct 04 05:10:19 crc kubenswrapper[4575]: I1004 05:10:19.983370 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerName="extract-utilities" Oct 04 05:10:19 crc kubenswrapper[4575]: E1004 05:10:19.983389 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerName="extract-content" Oct 04 05:10:19 crc kubenswrapper[4575]: I1004 05:10:19.983397 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerName="extract-content" Oct 04 05:10:19 crc kubenswrapper[4575]: E1004 05:10:19.983436 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerName="registry-server" Oct 04 05:10:19 crc kubenswrapper[4575]: I1004 05:10:19.983443 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerName="registry-server" Oct 04 05:10:19 crc kubenswrapper[4575]: I1004 05:10:19.983679 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5e29924-7c6e-428c-b8a2-8e9c4e9cdc2c" containerName="registry-server" Oct 04 05:10:19 crc kubenswrapper[4575]: I1004 05:10:19.985298 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.006301 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7f2jb"] Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.159364 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-catalog-content\") pod \"redhat-operators-7f2jb\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.159464 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-utilities\") pod \"redhat-operators-7f2jb\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.159763 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmnn5\" (UniqueName: \"kubernetes.io/projected/af35e5cb-d753-4b1e-ba5e-c30dffe72427-kube-api-access-vmnn5\") pod \"redhat-operators-7f2jb\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.261820 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-catalog-content\") pod \"redhat-operators-7f2jb\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.261895 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-utilities\") pod \"redhat-operators-7f2jb\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.261943 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmnn5\" (UniqueName: \"kubernetes.io/projected/af35e5cb-d753-4b1e-ba5e-c30dffe72427-kube-api-access-vmnn5\") pod \"redhat-operators-7f2jb\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.262553 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-catalog-content\") pod \"redhat-operators-7f2jb\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.262659 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-utilities\") pod \"redhat-operators-7f2jb\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.284881 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmnn5\" (UniqueName: \"kubernetes.io/projected/af35e5cb-d753-4b1e-ba5e-c30dffe72427-kube-api-access-vmnn5\") pod \"redhat-operators-7f2jb\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.305539 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.875646 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7f2jb"] Oct 04 05:10:20 crc kubenswrapper[4575]: I1004 05:10:20.970098 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f2jb" event={"ID":"af35e5cb-d753-4b1e-ba5e-c30dffe72427","Type":"ContainerStarted","Data":"2bc51b51d542ff8d022d87f05117c2494057bba3d8e3d19a82d287e3cf069d41"} Oct 04 05:10:21 crc kubenswrapper[4575]: I1004 05:10:21.980311 4575 generic.go:334] "Generic (PLEG): container finished" podID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerID="424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf" exitCode=0 Oct 04 05:10:21 crc kubenswrapper[4575]: I1004 05:10:21.980361 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f2jb" event={"ID":"af35e5cb-d753-4b1e-ba5e-c30dffe72427","Type":"ContainerDied","Data":"424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf"} Oct 04 05:10:24 crc kubenswrapper[4575]: I1004 05:10:24.000631 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f2jb" event={"ID":"af35e5cb-d753-4b1e-ba5e-c30dffe72427","Type":"ContainerStarted","Data":"b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f"} Oct 04 05:10:30 crc kubenswrapper[4575]: I1004 05:10:30.412834 4575 scope.go:117] "RemoveContainer" containerID="dd570d023229065a4fc9fdafdb94922c41e48208913becd257a5357f7f92c95c" Oct 04 05:10:32 crc kubenswrapper[4575]: I1004 05:10:32.069327 4575 generic.go:334] "Generic (PLEG): container finished" podID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerID="b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f" exitCode=0 Oct 04 05:10:32 crc kubenswrapper[4575]: I1004 05:10:32.069448 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f2jb" event={"ID":"af35e5cb-d753-4b1e-ba5e-c30dffe72427","Type":"ContainerDied","Data":"b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f"} Oct 04 05:10:33 crc kubenswrapper[4575]: I1004 05:10:33.081713 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f2jb" event={"ID":"af35e5cb-d753-4b1e-ba5e-c30dffe72427","Type":"ContainerStarted","Data":"5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69"} Oct 04 05:10:33 crc kubenswrapper[4575]: I1004 05:10:33.109702 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7f2jb" podStartSLOduration=3.307811616 podStartE2EDuration="14.109679531s" podCreationTimestamp="2025-10-04 05:10:19 +0000 UTC" firstStartedPulling="2025-10-04 05:10:21.983820224 +0000 UTC m=+2173.312379028" lastFinishedPulling="2025-10-04 05:10:32.785688129 +0000 UTC m=+2184.114246943" observedRunningTime="2025-10-04 05:10:33.101930196 +0000 UTC m=+2184.430489010" watchObservedRunningTime="2025-10-04 05:10:33.109679531 +0000 UTC m=+2184.438238355" Oct 04 05:10:35 crc kubenswrapper[4575]: I1004 05:10:35.097107 4575 generic.go:334] "Generic (PLEG): container finished" podID="5734782c-1920-445b-9225-f3172b8f28c8" containerID="758d40648a3c98d2f16d6f77edcec7ef8da26a439914f24a1674cf2d689e5255" exitCode=0 Oct 04 05:10:35 crc kubenswrapper[4575]: I1004 05:10:35.097419 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" event={"ID":"5734782c-1920-445b-9225-f3172b8f28c8","Type":"ContainerDied","Data":"758d40648a3c98d2f16d6f77edcec7ef8da26a439914f24a1674cf2d689e5255"} Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.696162 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.885472 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-ssh-key\") pod \"5734782c-1920-445b-9225-f3172b8f28c8\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.885847 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-inventory\") pod \"5734782c-1920-445b-9225-f3172b8f28c8\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.885957 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvzcg\" (UniqueName: \"kubernetes.io/projected/5734782c-1920-445b-9225-f3172b8f28c8-kube-api-access-rvzcg\") pod \"5734782c-1920-445b-9225-f3172b8f28c8\" (UID: \"5734782c-1920-445b-9225-f3172b8f28c8\") " Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.891902 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5734782c-1920-445b-9225-f3172b8f28c8-kube-api-access-rvzcg" (OuterVolumeSpecName: "kube-api-access-rvzcg") pod "5734782c-1920-445b-9225-f3172b8f28c8" (UID: "5734782c-1920-445b-9225-f3172b8f28c8"). InnerVolumeSpecName "kube-api-access-rvzcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.915626 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-inventory" (OuterVolumeSpecName: "inventory") pod "5734782c-1920-445b-9225-f3172b8f28c8" (UID: "5734782c-1920-445b-9225-f3172b8f28c8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.915790 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5734782c-1920-445b-9225-f3172b8f28c8" (UID: "5734782c-1920-445b-9225-f3172b8f28c8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.988984 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.989037 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvzcg\" (UniqueName: \"kubernetes.io/projected/5734782c-1920-445b-9225-f3172b8f28c8-kube-api-access-rvzcg\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:36 crc kubenswrapper[4575]: I1004 05:10:36.989050 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5734782c-1920-445b-9225-f3172b8f28c8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.113986 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" event={"ID":"5734782c-1920-445b-9225-f3172b8f28c8","Type":"ContainerDied","Data":"2868a8bf0d04bff69c7ebd1042a90111230a5c1e6f8994434ecb1c406334cdba"} Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.114027 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2868a8bf0d04bff69c7ebd1042a90111230a5c1e6f8994434ecb1c406334cdba" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.114063 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6gfmr" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.232386 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2"] Oct 04 05:10:37 crc kubenswrapper[4575]: E1004 05:10:37.233055 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5734782c-1920-445b-9225-f3172b8f28c8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.233084 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="5734782c-1920-445b-9225-f3172b8f28c8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.233344 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="5734782c-1920-445b-9225-f3172b8f28c8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.234294 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.239983 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.240174 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.240322 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.240498 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.258478 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2"] Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.396126 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-26wq2\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.396273 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbq6k\" (UniqueName: \"kubernetes.io/projected/cf592400-cd33-402d-b53f-6e378a7636cc-kube-api-access-kbq6k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-26wq2\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.396722 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-26wq2\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.500203 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-26wq2\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.500958 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-26wq2\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.501005 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbq6k\" (UniqueName: \"kubernetes.io/projected/cf592400-cd33-402d-b53f-6e378a7636cc-kube-api-access-kbq6k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-26wq2\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.508515 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-26wq2\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.515122 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-26wq2\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.520137 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbq6k\" (UniqueName: \"kubernetes.io/projected/cf592400-cd33-402d-b53f-6e378a7636cc-kube-api-access-kbq6k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-26wq2\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:37 crc kubenswrapper[4575]: I1004 05:10:37.555798 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:10:38 crc kubenswrapper[4575]: W1004 05:10:38.152610 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf592400_cd33_402d_b53f_6e378a7636cc.slice/crio-3cd77f9b40e759d2d3e71113c3137409a280f23188e20296dcaf9daf33a87172 WatchSource:0}: Error finding container 3cd77f9b40e759d2d3e71113c3137409a280f23188e20296dcaf9daf33a87172: Status 404 returned error can't find the container with id 3cd77f9b40e759d2d3e71113c3137409a280f23188e20296dcaf9daf33a87172 Oct 04 05:10:38 crc kubenswrapper[4575]: I1004 05:10:38.166082 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2"] Oct 04 05:10:39 crc kubenswrapper[4575]: I1004 05:10:39.134834 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" event={"ID":"cf592400-cd33-402d-b53f-6e378a7636cc","Type":"ContainerStarted","Data":"3cd77f9b40e759d2d3e71113c3137409a280f23188e20296dcaf9daf33a87172"} Oct 04 05:10:40 crc kubenswrapper[4575]: I1004 05:10:40.153370 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" event={"ID":"cf592400-cd33-402d-b53f-6e378a7636cc","Type":"ContainerStarted","Data":"ffc1cf8956fab232f91ed98952209ffff18c6f1fa9251adc218b2d269ae227f4"} Oct 04 05:10:40 crc kubenswrapper[4575]: I1004 05:10:40.177371 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" podStartSLOduration=2.22236313 podStartE2EDuration="3.177351503s" podCreationTimestamp="2025-10-04 05:10:37 +0000 UTC" firstStartedPulling="2025-10-04 05:10:38.154343679 +0000 UTC m=+2189.482902493" lastFinishedPulling="2025-10-04 05:10:39.109332062 +0000 UTC m=+2190.437890866" observedRunningTime="2025-10-04 05:10:40.17173564 +0000 UTC m=+2191.500294454" watchObservedRunningTime="2025-10-04 05:10:40.177351503 +0000 UTC m=+2191.505910317" Oct 04 05:10:40 crc kubenswrapper[4575]: I1004 05:10:40.306425 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:40 crc kubenswrapper[4575]: I1004 05:10:40.306728 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:40 crc kubenswrapper[4575]: I1004 05:10:40.360615 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:41 crc kubenswrapper[4575]: I1004 05:10:41.210289 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:41 crc kubenswrapper[4575]: I1004 05:10:41.260526 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7f2jb"] Oct 04 05:10:43 crc kubenswrapper[4575]: I1004 05:10:43.177934 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7f2jb" podUID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerName="registry-server" containerID="cri-o://5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69" gracePeriod=2 Oct 04 05:10:43 crc kubenswrapper[4575]: I1004 05:10:43.805402 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:43 crc kubenswrapper[4575]: I1004 05:10:43.926645 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-utilities\") pod \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " Oct 04 05:10:43 crc kubenswrapper[4575]: I1004 05:10:43.927068 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-catalog-content\") pod \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " Oct 04 05:10:43 crc kubenswrapper[4575]: I1004 05:10:43.927199 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmnn5\" (UniqueName: \"kubernetes.io/projected/af35e5cb-d753-4b1e-ba5e-c30dffe72427-kube-api-access-vmnn5\") pod \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\" (UID: \"af35e5cb-d753-4b1e-ba5e-c30dffe72427\") " Oct 04 05:10:43 crc kubenswrapper[4575]: I1004 05:10:43.927921 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-utilities" (OuterVolumeSpecName: "utilities") pod "af35e5cb-d753-4b1e-ba5e-c30dffe72427" (UID: "af35e5cb-d753-4b1e-ba5e-c30dffe72427"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:43 crc kubenswrapper[4575]: I1004 05:10:43.933520 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af35e5cb-d753-4b1e-ba5e-c30dffe72427-kube-api-access-vmnn5" (OuterVolumeSpecName: "kube-api-access-vmnn5") pod "af35e5cb-d753-4b1e-ba5e-c30dffe72427" (UID: "af35e5cb-d753-4b1e-ba5e-c30dffe72427"). InnerVolumeSpecName "kube-api-access-vmnn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.029157 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmnn5\" (UniqueName: \"kubernetes.io/projected/af35e5cb-d753-4b1e-ba5e-c30dffe72427-kube-api-access-vmnn5\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.029200 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.029305 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af35e5cb-d753-4b1e-ba5e-c30dffe72427" (UID: "af35e5cb-d753-4b1e-ba5e-c30dffe72427"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.132276 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af35e5cb-d753-4b1e-ba5e-c30dffe72427-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.195956 4575 generic.go:334] "Generic (PLEG): container finished" podID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerID="5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69" exitCode=0 Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.197008 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f2jb" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.197049 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f2jb" event={"ID":"af35e5cb-d753-4b1e-ba5e-c30dffe72427","Type":"ContainerDied","Data":"5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69"} Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.199329 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f2jb" event={"ID":"af35e5cb-d753-4b1e-ba5e-c30dffe72427","Type":"ContainerDied","Data":"2bc51b51d542ff8d022d87f05117c2494057bba3d8e3d19a82d287e3cf069d41"} Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.199366 4575 scope.go:117] "RemoveContainer" containerID="5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.230976 4575 scope.go:117] "RemoveContainer" containerID="b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.241404 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7f2jb"] Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.253098 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7f2jb"] Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.272224 4575 scope.go:117] "RemoveContainer" containerID="424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.331871 4575 scope.go:117] "RemoveContainer" containerID="5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69" Oct 04 05:10:44 crc kubenswrapper[4575]: E1004 05:10:44.333216 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69\": container with ID starting with 5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69 not found: ID does not exist" containerID="5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.333252 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69"} err="failed to get container status \"5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69\": rpc error: code = NotFound desc = could not find container \"5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69\": container with ID starting with 5d212a9b9be23de621fd36d070637eeb11d8def619fbfd6e50ab30b8f4f64b69 not found: ID does not exist" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.333280 4575 scope.go:117] "RemoveContainer" containerID="b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f" Oct 04 05:10:44 crc kubenswrapper[4575]: E1004 05:10:44.333906 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f\": container with ID starting with b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f not found: ID does not exist" containerID="b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.333940 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f"} err="failed to get container status \"b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f\": rpc error: code = NotFound desc = could not find container \"b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f\": container with ID starting with b57df3294987fa72bc8dcce1102ace146fc95673a1ab07ae696b3047b5c4899f not found: ID does not exist" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.333959 4575 scope.go:117] "RemoveContainer" containerID="424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf" Oct 04 05:10:44 crc kubenswrapper[4575]: E1004 05:10:44.334303 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf\": container with ID starting with 424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf not found: ID does not exist" containerID="424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf" Oct 04 05:10:44 crc kubenswrapper[4575]: I1004 05:10:44.334334 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf"} err="failed to get container status \"424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf\": rpc error: code = NotFound desc = could not find container \"424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf\": container with ID starting with 424b15eec34abafdc473a6878581bb4b63369fb6d863405e7db745b9369b7bcf not found: ID does not exist" Oct 04 05:10:45 crc kubenswrapper[4575]: I1004 05:10:45.325178 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" path="/var/lib/kubelet/pods/af35e5cb-d753-4b1e-ba5e-c30dffe72427/volumes" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.492133 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mn94q"] Oct 04 05:11:05 crc kubenswrapper[4575]: E1004 05:11:05.493098 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerName="extract-content" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.493114 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerName="extract-content" Oct 04 05:11:05 crc kubenswrapper[4575]: E1004 05:11:05.493129 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerName="registry-server" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.493137 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerName="registry-server" Oct 04 05:11:05 crc kubenswrapper[4575]: E1004 05:11:05.493172 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerName="extract-utilities" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.493179 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerName="extract-utilities" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.493426 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="af35e5cb-d753-4b1e-ba5e-c30dffe72427" containerName="registry-server" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.495307 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.514563 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mn94q"] Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.570985 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-utilities\") pod \"community-operators-mn94q\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.571063 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-catalog-content\") pod \"community-operators-mn94q\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.571431 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xn5q\" (UniqueName: \"kubernetes.io/projected/f2421c97-a04b-48da-8368-6f203518e22a-kube-api-access-8xn5q\") pod \"community-operators-mn94q\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.673294 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-utilities\") pod \"community-operators-mn94q\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.673362 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-catalog-content\") pod \"community-operators-mn94q\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.673485 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xn5q\" (UniqueName: \"kubernetes.io/projected/f2421c97-a04b-48da-8368-6f203518e22a-kube-api-access-8xn5q\") pod \"community-operators-mn94q\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.673818 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-utilities\") pod \"community-operators-mn94q\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.674194 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-catalog-content\") pod \"community-operators-mn94q\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.699410 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xn5q\" (UniqueName: \"kubernetes.io/projected/f2421c97-a04b-48da-8368-6f203518e22a-kube-api-access-8xn5q\") pod \"community-operators-mn94q\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:05 crc kubenswrapper[4575]: I1004 05:11:05.821189 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:06 crc kubenswrapper[4575]: I1004 05:11:06.212148 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mn94q"] Oct 04 05:11:06 crc kubenswrapper[4575]: I1004 05:11:06.383881 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mn94q" event={"ID":"f2421c97-a04b-48da-8368-6f203518e22a","Type":"ContainerStarted","Data":"3b70ecb23e9d00bbeebef19f3e36deb2a0af8733d4782054fdd6efb4fa559ba5"} Oct 04 05:11:07 crc kubenswrapper[4575]: I1004 05:11:07.403974 4575 generic.go:334] "Generic (PLEG): container finished" podID="f2421c97-a04b-48da-8368-6f203518e22a" containerID="3fbf183b2da0d476bc5f000443c3578c99fca27b6bf7aa43434c6a495dfccd35" exitCode=0 Oct 04 05:11:07 crc kubenswrapper[4575]: I1004 05:11:07.404988 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mn94q" event={"ID":"f2421c97-a04b-48da-8368-6f203518e22a","Type":"ContainerDied","Data":"3fbf183b2da0d476bc5f000443c3578c99fca27b6bf7aa43434c6a495dfccd35"} Oct 04 05:11:09 crc kubenswrapper[4575]: I1004 05:11:09.423693 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mn94q" event={"ID":"f2421c97-a04b-48da-8368-6f203518e22a","Type":"ContainerStarted","Data":"a60d3ac10e770ebf1bdbd094f9f953ca7eb2787fb381e4c9635f922bb7c2499c"} Oct 04 05:11:12 crc kubenswrapper[4575]: I1004 05:11:12.450994 4575 generic.go:334] "Generic (PLEG): container finished" podID="f2421c97-a04b-48da-8368-6f203518e22a" containerID="a60d3ac10e770ebf1bdbd094f9f953ca7eb2787fb381e4c9635f922bb7c2499c" exitCode=0 Oct 04 05:11:12 crc kubenswrapper[4575]: I1004 05:11:12.451082 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mn94q" event={"ID":"f2421c97-a04b-48da-8368-6f203518e22a","Type":"ContainerDied","Data":"a60d3ac10e770ebf1bdbd094f9f953ca7eb2787fb381e4c9635f922bb7c2499c"} Oct 04 05:11:15 crc kubenswrapper[4575]: I1004 05:11:15.478487 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mn94q" event={"ID":"f2421c97-a04b-48da-8368-6f203518e22a","Type":"ContainerStarted","Data":"450af5c33f2eea3c34a197e3f202924e6240c60880bb5c8d23dccc8c71cc89cb"} Oct 04 05:11:15 crc kubenswrapper[4575]: I1004 05:11:15.501281 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mn94q" podStartSLOduration=3.308071737 podStartE2EDuration="10.501259857s" podCreationTimestamp="2025-10-04 05:11:05 +0000 UTC" firstStartedPulling="2025-10-04 05:11:07.408433052 +0000 UTC m=+2218.736991866" lastFinishedPulling="2025-10-04 05:11:14.601621172 +0000 UTC m=+2225.930179986" observedRunningTime="2025-10-04 05:11:15.493738409 +0000 UTC m=+2226.822297223" watchObservedRunningTime="2025-10-04 05:11:15.501259857 +0000 UTC m=+2226.829818681" Oct 04 05:11:15 crc kubenswrapper[4575]: I1004 05:11:15.822192 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:15 crc kubenswrapper[4575]: I1004 05:11:15.822262 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:16 crc kubenswrapper[4575]: I1004 05:11:16.879415 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mn94q" podUID="f2421c97-a04b-48da-8368-6f203518e22a" containerName="registry-server" probeResult="failure" output=< Oct 04 05:11:16 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:11:16 crc kubenswrapper[4575]: > Oct 04 05:11:25 crc kubenswrapper[4575]: I1004 05:11:25.873693 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:25 crc kubenswrapper[4575]: I1004 05:11:25.943489 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:26 crc kubenswrapper[4575]: I1004 05:11:26.108842 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mn94q"] Oct 04 05:11:27 crc kubenswrapper[4575]: I1004 05:11:27.580496 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mn94q" podUID="f2421c97-a04b-48da-8368-6f203518e22a" containerName="registry-server" containerID="cri-o://450af5c33f2eea3c34a197e3f202924e6240c60880bb5c8d23dccc8c71cc89cb" gracePeriod=2 Oct 04 05:11:28 crc kubenswrapper[4575]: I1004 05:11:28.332784 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-9w4x4" podUID="27a10d08-b3eb-4d6d-8815-9db3a56313c5" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 05:11:28 crc kubenswrapper[4575]: I1004 05:11:28.593885 4575 generic.go:334] "Generic (PLEG): container finished" podID="f2421c97-a04b-48da-8368-6f203518e22a" containerID="450af5c33f2eea3c34a197e3f202924e6240c60880bb5c8d23dccc8c71cc89cb" exitCode=0 Oct 04 05:11:28 crc kubenswrapper[4575]: I1004 05:11:28.593965 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mn94q" event={"ID":"f2421c97-a04b-48da-8368-6f203518e22a","Type":"ContainerDied","Data":"450af5c33f2eea3c34a197e3f202924e6240c60880bb5c8d23dccc8c71cc89cb"} Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.011466 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.146914 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-catalog-content\") pod \"f2421c97-a04b-48da-8368-6f203518e22a\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.147074 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-utilities\") pod \"f2421c97-a04b-48da-8368-6f203518e22a\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.147200 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xn5q\" (UniqueName: \"kubernetes.io/projected/f2421c97-a04b-48da-8368-6f203518e22a-kube-api-access-8xn5q\") pod \"f2421c97-a04b-48da-8368-6f203518e22a\" (UID: \"f2421c97-a04b-48da-8368-6f203518e22a\") " Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.149425 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-utilities" (OuterVolumeSpecName: "utilities") pod "f2421c97-a04b-48da-8368-6f203518e22a" (UID: "f2421c97-a04b-48da-8368-6f203518e22a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.154268 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2421c97-a04b-48da-8368-6f203518e22a-kube-api-access-8xn5q" (OuterVolumeSpecName: "kube-api-access-8xn5q") pod "f2421c97-a04b-48da-8368-6f203518e22a" (UID: "f2421c97-a04b-48da-8368-6f203518e22a"). InnerVolumeSpecName "kube-api-access-8xn5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.201485 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f2421c97-a04b-48da-8368-6f203518e22a" (UID: "f2421c97-a04b-48da-8368-6f203518e22a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.250038 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.250302 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xn5q\" (UniqueName: \"kubernetes.io/projected/f2421c97-a04b-48da-8368-6f203518e22a-kube-api-access-8xn5q\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.250314 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2421c97-a04b-48da-8368-6f203518e22a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.605301 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mn94q" event={"ID":"f2421c97-a04b-48da-8368-6f203518e22a","Type":"ContainerDied","Data":"3b70ecb23e9d00bbeebef19f3e36deb2a0af8733d4782054fdd6efb4fa559ba5"} Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.605368 4575 scope.go:117] "RemoveContainer" containerID="450af5c33f2eea3c34a197e3f202924e6240c60880bb5c8d23dccc8c71cc89cb" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.605737 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mn94q" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.638823 4575 scope.go:117] "RemoveContainer" containerID="a60d3ac10e770ebf1bdbd094f9f953ca7eb2787fb381e4c9635f922bb7c2499c" Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.642187 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mn94q"] Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.660415 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mn94q"] Oct 04 05:11:29 crc kubenswrapper[4575]: I1004 05:11:29.665650 4575 scope.go:117] "RemoveContainer" containerID="3fbf183b2da0d476bc5f000443c3578c99fca27b6bf7aa43434c6a495dfccd35" Oct 04 05:11:31 crc kubenswrapper[4575]: I1004 05:11:31.322406 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2421c97-a04b-48da-8368-6f203518e22a" path="/var/lib/kubelet/pods/f2421c97-a04b-48da-8368-6f203518e22a/volumes" Oct 04 05:11:37 crc kubenswrapper[4575]: I1004 05:11:37.671626 4575 generic.go:334] "Generic (PLEG): container finished" podID="cf592400-cd33-402d-b53f-6e378a7636cc" containerID="ffc1cf8956fab232f91ed98952209ffff18c6f1fa9251adc218b2d269ae227f4" exitCode=2 Oct 04 05:11:37 crc kubenswrapper[4575]: I1004 05:11:37.672099 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" event={"ID":"cf592400-cd33-402d-b53f-6e378a7636cc","Type":"ContainerDied","Data":"ffc1cf8956fab232f91ed98952209ffff18c6f1fa9251adc218b2d269ae227f4"} Oct 04 05:11:38 crc kubenswrapper[4575]: I1004 05:11:38.446670 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:11:38 crc kubenswrapper[4575]: I1004 05:11:38.446730 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.150859 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.244413 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbq6k\" (UniqueName: \"kubernetes.io/projected/cf592400-cd33-402d-b53f-6e378a7636cc-kube-api-access-kbq6k\") pod \"cf592400-cd33-402d-b53f-6e378a7636cc\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.244728 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-inventory\") pod \"cf592400-cd33-402d-b53f-6e378a7636cc\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.244801 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-ssh-key\") pod \"cf592400-cd33-402d-b53f-6e378a7636cc\" (UID: \"cf592400-cd33-402d-b53f-6e378a7636cc\") " Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.252029 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf592400-cd33-402d-b53f-6e378a7636cc-kube-api-access-kbq6k" (OuterVolumeSpecName: "kube-api-access-kbq6k") pod "cf592400-cd33-402d-b53f-6e378a7636cc" (UID: "cf592400-cd33-402d-b53f-6e378a7636cc"). InnerVolumeSpecName "kube-api-access-kbq6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.274055 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cf592400-cd33-402d-b53f-6e378a7636cc" (UID: "cf592400-cd33-402d-b53f-6e378a7636cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.276324 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-inventory" (OuterVolumeSpecName: "inventory") pod "cf592400-cd33-402d-b53f-6e378a7636cc" (UID: "cf592400-cd33-402d-b53f-6e378a7636cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.347252 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbq6k\" (UniqueName: \"kubernetes.io/projected/cf592400-cd33-402d-b53f-6e378a7636cc-kube-api-access-kbq6k\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.347290 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.347302 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf592400-cd33-402d-b53f-6e378a7636cc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.697493 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" event={"ID":"cf592400-cd33-402d-b53f-6e378a7636cc","Type":"ContainerDied","Data":"3cd77f9b40e759d2d3e71113c3137409a280f23188e20296dcaf9daf33a87172"} Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.697536 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cd77f9b40e759d2d3e71113c3137409a280f23188e20296dcaf9daf33a87172" Oct 04 05:11:39 crc kubenswrapper[4575]: I1004 05:11:39.697625 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-26wq2" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.047844 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zgbhz"] Oct 04 05:11:41 crc kubenswrapper[4575]: E1004 05:11:41.048579 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2421c97-a04b-48da-8368-6f203518e22a" containerName="extract-utilities" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.048611 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2421c97-a04b-48da-8368-6f203518e22a" containerName="extract-utilities" Oct 04 05:11:41 crc kubenswrapper[4575]: E1004 05:11:41.048619 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2421c97-a04b-48da-8368-6f203518e22a" containerName="extract-content" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.048627 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2421c97-a04b-48da-8368-6f203518e22a" containerName="extract-content" Oct 04 05:11:41 crc kubenswrapper[4575]: E1004 05:11:41.048644 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2421c97-a04b-48da-8368-6f203518e22a" containerName="registry-server" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.048653 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2421c97-a04b-48da-8368-6f203518e22a" containerName="registry-server" Oct 04 05:11:41 crc kubenswrapper[4575]: E1004 05:11:41.048665 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf592400-cd33-402d-b53f-6e378a7636cc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.048671 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf592400-cd33-402d-b53f-6e378a7636cc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.048890 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2421c97-a04b-48da-8368-6f203518e22a" containerName="registry-server" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.048909 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf592400-cd33-402d-b53f-6e378a7636cc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.050226 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.085706 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgbhz"] Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.189849 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-catalog-content\") pod \"redhat-marketplace-zgbhz\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.190122 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6gpq\" (UniqueName: \"kubernetes.io/projected/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-kube-api-access-d6gpq\") pod \"redhat-marketplace-zgbhz\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.190320 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-utilities\") pod \"redhat-marketplace-zgbhz\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.292329 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-catalog-content\") pod \"redhat-marketplace-zgbhz\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.292964 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6gpq\" (UniqueName: \"kubernetes.io/projected/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-kube-api-access-d6gpq\") pod \"redhat-marketplace-zgbhz\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.293026 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-utilities\") pod \"redhat-marketplace-zgbhz\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.292886 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-catalog-content\") pod \"redhat-marketplace-zgbhz\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.293473 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-utilities\") pod \"redhat-marketplace-zgbhz\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.315192 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6gpq\" (UniqueName: \"kubernetes.io/projected/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-kube-api-access-d6gpq\") pod \"redhat-marketplace-zgbhz\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.370488 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:41 crc kubenswrapper[4575]: I1004 05:11:41.861442 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgbhz"] Oct 04 05:11:42 crc kubenswrapper[4575]: I1004 05:11:42.722267 4575 generic.go:334] "Generic (PLEG): container finished" podID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerID="ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd" exitCode=0 Oct 04 05:11:42 crc kubenswrapper[4575]: I1004 05:11:42.722364 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgbhz" event={"ID":"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef","Type":"ContainerDied","Data":"ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd"} Oct 04 05:11:42 crc kubenswrapper[4575]: I1004 05:11:42.722577 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgbhz" event={"ID":"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef","Type":"ContainerStarted","Data":"d822e2acff144f315ad1fa7a0c05254641354922c52a21d79718d6f8be8bd780"} Oct 04 05:11:42 crc kubenswrapper[4575]: I1004 05:11:42.727168 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:11:44 crc kubenswrapper[4575]: I1004 05:11:44.742821 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgbhz" event={"ID":"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef","Type":"ContainerStarted","Data":"f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b"} Oct 04 05:11:45 crc kubenswrapper[4575]: I1004 05:11:45.752984 4575 generic.go:334] "Generic (PLEG): container finished" podID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerID="f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b" exitCode=0 Oct 04 05:11:45 crc kubenswrapper[4575]: I1004 05:11:45.753076 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgbhz" event={"ID":"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef","Type":"ContainerDied","Data":"f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b"} Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.032016 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7"] Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.033700 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.039435 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.039458 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.039696 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.039641 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.041035 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7"] Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.117002 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnqcv\" (UniqueName: \"kubernetes.io/projected/2fed43b6-f18d-44d6-85f9-c543c5b81935-kube-api-access-hnqcv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.117333 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.117664 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.219553 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.219679 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.219800 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnqcv\" (UniqueName: \"kubernetes.io/projected/2fed43b6-f18d-44d6-85f9-c543c5b81935-kube-api-access-hnqcv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.225359 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.238265 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.243781 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnqcv\" (UniqueName: \"kubernetes.io/projected/2fed43b6-f18d-44d6-85f9-c543c5b81935-kube-api-access-hnqcv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.351631 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.786135 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgbhz" event={"ID":"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef","Type":"ContainerStarted","Data":"0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597"} Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.808251 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zgbhz" podStartSLOduration=2.816983618 podStartE2EDuration="6.808231708s" podCreationTimestamp="2025-10-04 05:11:41 +0000 UTC" firstStartedPulling="2025-10-04 05:11:42.726892429 +0000 UTC m=+2254.055451243" lastFinishedPulling="2025-10-04 05:11:46.718140519 +0000 UTC m=+2258.046699333" observedRunningTime="2025-10-04 05:11:47.801931026 +0000 UTC m=+2259.130489840" watchObservedRunningTime="2025-10-04 05:11:47.808231708 +0000 UTC m=+2259.136790532" Oct 04 05:11:47 crc kubenswrapper[4575]: W1004 05:11:47.907948 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fed43b6_f18d_44d6_85f9_c543c5b81935.slice/crio-66b1cd5c7d310ea0b69519b2ccef162ddaf01188a859ba392df86ede251bae01 WatchSource:0}: Error finding container 66b1cd5c7d310ea0b69519b2ccef162ddaf01188a859ba392df86ede251bae01: Status 404 returned error can't find the container with id 66b1cd5c7d310ea0b69519b2ccef162ddaf01188a859ba392df86ede251bae01 Oct 04 05:11:47 crc kubenswrapper[4575]: I1004 05:11:47.908113 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7"] Oct 04 05:11:48 crc kubenswrapper[4575]: I1004 05:11:48.802523 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" event={"ID":"2fed43b6-f18d-44d6-85f9-c543c5b81935","Type":"ContainerStarted","Data":"0e7c559fd2497c82df6fd5b8476d82907c4d3b161f038a679a99a1a58b333e8a"} Oct 04 05:11:48 crc kubenswrapper[4575]: I1004 05:11:48.803020 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" event={"ID":"2fed43b6-f18d-44d6-85f9-c543c5b81935","Type":"ContainerStarted","Data":"66b1cd5c7d310ea0b69519b2ccef162ddaf01188a859ba392df86ede251bae01"} Oct 04 05:11:48 crc kubenswrapper[4575]: I1004 05:11:48.823425 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" podStartSLOduration=1.394583924 podStartE2EDuration="1.823401989s" podCreationTimestamp="2025-10-04 05:11:47 +0000 UTC" firstStartedPulling="2025-10-04 05:11:47.911688083 +0000 UTC m=+2259.240246907" lastFinishedPulling="2025-10-04 05:11:48.340506168 +0000 UTC m=+2259.669064972" observedRunningTime="2025-10-04 05:11:48.816222731 +0000 UTC m=+2260.144781565" watchObservedRunningTime="2025-10-04 05:11:48.823401989 +0000 UTC m=+2260.151960803" Oct 04 05:11:51 crc kubenswrapper[4575]: I1004 05:11:51.372741 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:51 crc kubenswrapper[4575]: I1004 05:11:51.373312 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:51 crc kubenswrapper[4575]: I1004 05:11:51.423635 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:51 crc kubenswrapper[4575]: I1004 05:11:51.876860 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:51 crc kubenswrapper[4575]: I1004 05:11:51.928926 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgbhz"] Oct 04 05:11:53 crc kubenswrapper[4575]: I1004 05:11:53.845233 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zgbhz" podUID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerName="registry-server" containerID="cri-o://0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597" gracePeriod=2 Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.353268 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.455436 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-utilities\") pod \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.455690 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6gpq\" (UniqueName: \"kubernetes.io/projected/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-kube-api-access-d6gpq\") pod \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.455849 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-catalog-content\") pod \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\" (UID: \"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef\") " Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.457455 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-utilities" (OuterVolumeSpecName: "utilities") pod "d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" (UID: "d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.461496 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-kube-api-access-d6gpq" (OuterVolumeSpecName: "kube-api-access-d6gpq") pod "d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" (UID: "d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef"). InnerVolumeSpecName "kube-api-access-d6gpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.470411 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" (UID: "d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.558699 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.558730 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6gpq\" (UniqueName: \"kubernetes.io/projected/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-kube-api-access-d6gpq\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.558741 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.856494 4575 generic.go:334] "Generic (PLEG): container finished" podID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerID="0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597" exitCode=0 Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.856551 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgbhz" event={"ID":"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef","Type":"ContainerDied","Data":"0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597"} Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.856563 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zgbhz" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.856604 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zgbhz" event={"ID":"d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef","Type":"ContainerDied","Data":"d822e2acff144f315ad1fa7a0c05254641354922c52a21d79718d6f8be8bd780"} Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.856628 4575 scope.go:117] "RemoveContainer" containerID="0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.881271 4575 scope.go:117] "RemoveContainer" containerID="f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.905522 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgbhz"] Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.911268 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zgbhz"] Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.912998 4575 scope.go:117] "RemoveContainer" containerID="ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.962056 4575 scope.go:117] "RemoveContainer" containerID="0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597" Oct 04 05:11:54 crc kubenswrapper[4575]: E1004 05:11:54.962640 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597\": container with ID starting with 0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597 not found: ID does not exist" containerID="0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.962679 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597"} err="failed to get container status \"0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597\": rpc error: code = NotFound desc = could not find container \"0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597\": container with ID starting with 0565e7b0a3cbb656d8327852487fd1ebf659a7b5d7383e1fc5c066aa8a35c597 not found: ID does not exist" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.962718 4575 scope.go:117] "RemoveContainer" containerID="f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b" Oct 04 05:11:54 crc kubenswrapper[4575]: E1004 05:11:54.963248 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b\": container with ID starting with f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b not found: ID does not exist" containerID="f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.963290 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b"} err="failed to get container status \"f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b\": rpc error: code = NotFound desc = could not find container \"f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b\": container with ID starting with f3497f08cd30d20e74c26ad863b25e6711b769ef8d1b106312a23fe1538d7f0b not found: ID does not exist" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.963305 4575 scope.go:117] "RemoveContainer" containerID="ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd" Oct 04 05:11:54 crc kubenswrapper[4575]: E1004 05:11:54.963793 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd\": container with ID starting with ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd not found: ID does not exist" containerID="ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd" Oct 04 05:11:54 crc kubenswrapper[4575]: I1004 05:11:54.963844 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd"} err="failed to get container status \"ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd\": rpc error: code = NotFound desc = could not find container \"ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd\": container with ID starting with ac3ae67e0003eaca35c951431d50f91aa1bea7c135cece13f0ab9c27bbc8b2cd not found: ID does not exist" Oct 04 05:11:55 crc kubenswrapper[4575]: I1004 05:11:55.323933 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" path="/var/lib/kubelet/pods/d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef/volumes" Oct 04 05:12:08 crc kubenswrapper[4575]: I1004 05:12:08.446241 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:12:08 crc kubenswrapper[4575]: I1004 05:12:08.446840 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:12:35 crc kubenswrapper[4575]: I1004 05:12:35.194712 4575 generic.go:334] "Generic (PLEG): container finished" podID="2fed43b6-f18d-44d6-85f9-c543c5b81935" containerID="0e7c559fd2497c82df6fd5b8476d82907c4d3b161f038a679a99a1a58b333e8a" exitCode=0 Oct 04 05:12:35 crc kubenswrapper[4575]: I1004 05:12:35.195222 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" event={"ID":"2fed43b6-f18d-44d6-85f9-c543c5b81935","Type":"ContainerDied","Data":"0e7c559fd2497c82df6fd5b8476d82907c4d3b161f038a679a99a1a58b333e8a"} Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.680760 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.801817 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-ssh-key\") pod \"2fed43b6-f18d-44d6-85f9-c543c5b81935\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.801948 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnqcv\" (UniqueName: \"kubernetes.io/projected/2fed43b6-f18d-44d6-85f9-c543c5b81935-kube-api-access-hnqcv\") pod \"2fed43b6-f18d-44d6-85f9-c543c5b81935\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.802056 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-inventory\") pod \"2fed43b6-f18d-44d6-85f9-c543c5b81935\" (UID: \"2fed43b6-f18d-44d6-85f9-c543c5b81935\") " Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.807822 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fed43b6-f18d-44d6-85f9-c543c5b81935-kube-api-access-hnqcv" (OuterVolumeSpecName: "kube-api-access-hnqcv") pod "2fed43b6-f18d-44d6-85f9-c543c5b81935" (UID: "2fed43b6-f18d-44d6-85f9-c543c5b81935"). InnerVolumeSpecName "kube-api-access-hnqcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.835276 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2fed43b6-f18d-44d6-85f9-c543c5b81935" (UID: "2fed43b6-f18d-44d6-85f9-c543c5b81935"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.838443 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-inventory" (OuterVolumeSpecName: "inventory") pod "2fed43b6-f18d-44d6-85f9-c543c5b81935" (UID: "2fed43b6-f18d-44d6-85f9-c543c5b81935"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.904677 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.904723 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fed43b6-f18d-44d6-85f9-c543c5b81935-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:12:36 crc kubenswrapper[4575]: I1004 05:12:36.904736 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnqcv\" (UniqueName: \"kubernetes.io/projected/2fed43b6-f18d-44d6-85f9-c543c5b81935-kube-api-access-hnqcv\") on node \"crc\" DevicePath \"\"" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.229325 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" event={"ID":"2fed43b6-f18d-44d6-85f9-c543c5b81935","Type":"ContainerDied","Data":"66b1cd5c7d310ea0b69519b2ccef162ddaf01188a859ba392df86ede251bae01"} Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.229679 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66b1cd5c7d310ea0b69519b2ccef162ddaf01188a859ba392df86ede251bae01" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.229644 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.339469 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-j4phd"] Oct 04 05:12:37 crc kubenswrapper[4575]: E1004 05:12:37.339874 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerName="extract-utilities" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.339897 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerName="extract-utilities" Oct 04 05:12:37 crc kubenswrapper[4575]: E1004 05:12:37.339920 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerName="extract-content" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.339928 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerName="extract-content" Oct 04 05:12:37 crc kubenswrapper[4575]: E1004 05:12:37.339950 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fed43b6-f18d-44d6-85f9-c543c5b81935" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.339967 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fed43b6-f18d-44d6-85f9-c543c5b81935" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:12:37 crc kubenswrapper[4575]: E1004 05:12:37.339981 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerName="registry-server" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.339988 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerName="registry-server" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.340205 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fed43b6-f18d-44d6-85f9-c543c5b81935" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.340226 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c9a5d0-15b4-4575-9dfb-5946edf1a7ef" containerName="registry-server" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.341027 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.345902 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.346136 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.346280 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.346443 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.353711 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-j4phd"] Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.518350 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-j4phd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.518629 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gg49\" (UniqueName: \"kubernetes.io/projected/eff0e85f-146c-4e9f-8109-853df39fb0cd-kube-api-access-9gg49\") pod \"ssh-known-hosts-edpm-deployment-j4phd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.518957 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-j4phd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.620206 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gg49\" (UniqueName: \"kubernetes.io/projected/eff0e85f-146c-4e9f-8109-853df39fb0cd-kube-api-access-9gg49\") pod \"ssh-known-hosts-edpm-deployment-j4phd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.620345 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-j4phd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.620406 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-j4phd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.624345 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-j4phd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.630164 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-j4phd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.640956 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gg49\" (UniqueName: \"kubernetes.io/projected/eff0e85f-146c-4e9f-8109-853df39fb0cd-kube-api-access-9gg49\") pod \"ssh-known-hosts-edpm-deployment-j4phd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:37 crc kubenswrapper[4575]: I1004 05:12:37.660429 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:38 crc kubenswrapper[4575]: I1004 05:12:38.209294 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-j4phd"] Oct 04 05:12:38 crc kubenswrapper[4575]: I1004 05:12:38.241788 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" event={"ID":"eff0e85f-146c-4e9f-8109-853df39fb0cd","Type":"ContainerStarted","Data":"eadfc33dc4eed1f6ccd3a1875fafb77a5b3fcb267b38b5a9b379135f063ffd56"} Oct 04 05:12:38 crc kubenswrapper[4575]: I1004 05:12:38.446048 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:12:38 crc kubenswrapper[4575]: I1004 05:12:38.446454 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:12:38 crc kubenswrapper[4575]: I1004 05:12:38.446514 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:12:38 crc kubenswrapper[4575]: I1004 05:12:38.447390 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:12:38 crc kubenswrapper[4575]: I1004 05:12:38.447496 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" gracePeriod=600 Oct 04 05:12:38 crc kubenswrapper[4575]: E1004 05:12:38.590206 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:12:39 crc kubenswrapper[4575]: I1004 05:12:39.258820 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" exitCode=0 Oct 04 05:12:39 crc kubenswrapper[4575]: I1004 05:12:39.259022 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e"} Oct 04 05:12:39 crc kubenswrapper[4575]: I1004 05:12:39.259144 4575 scope.go:117] "RemoveContainer" containerID="ecff37b59205a74250a64b395a45822895a4e3324eeb81b77ca18757cc564c58" Oct 04 05:12:39 crc kubenswrapper[4575]: I1004 05:12:39.259828 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:12:39 crc kubenswrapper[4575]: E1004 05:12:39.260092 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:12:40 crc kubenswrapper[4575]: I1004 05:12:40.274998 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" event={"ID":"eff0e85f-146c-4e9f-8109-853df39fb0cd","Type":"ContainerStarted","Data":"77645c86683c56082d9d3a79c5af05f45e550f7fc27527b608cb7a997e342110"} Oct 04 05:12:40 crc kubenswrapper[4575]: I1004 05:12:40.299674 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" podStartSLOduration=2.581886113 podStartE2EDuration="3.299652603s" podCreationTimestamp="2025-10-04 05:12:37 +0000 UTC" firstStartedPulling="2025-10-04 05:12:38.220347754 +0000 UTC m=+2309.548906578" lastFinishedPulling="2025-10-04 05:12:38.938114254 +0000 UTC m=+2310.266673068" observedRunningTime="2025-10-04 05:12:40.297881522 +0000 UTC m=+2311.626440366" watchObservedRunningTime="2025-10-04 05:12:40.299652603 +0000 UTC m=+2311.628211417" Oct 04 05:12:47 crc kubenswrapper[4575]: I1004 05:12:47.362568 4575 generic.go:334] "Generic (PLEG): container finished" podID="eff0e85f-146c-4e9f-8109-853df39fb0cd" containerID="77645c86683c56082d9d3a79c5af05f45e550f7fc27527b608cb7a997e342110" exitCode=0 Oct 04 05:12:47 crc kubenswrapper[4575]: I1004 05:12:47.362643 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" event={"ID":"eff0e85f-146c-4e9f-8109-853df39fb0cd","Type":"ContainerDied","Data":"77645c86683c56082d9d3a79c5af05f45e550f7fc27527b608cb7a997e342110"} Oct 04 05:12:48 crc kubenswrapper[4575]: I1004 05:12:48.802675 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:48 crc kubenswrapper[4575]: I1004 05:12:48.940544 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gg49\" (UniqueName: \"kubernetes.io/projected/eff0e85f-146c-4e9f-8109-853df39fb0cd-kube-api-access-9gg49\") pod \"eff0e85f-146c-4e9f-8109-853df39fb0cd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " Oct 04 05:12:48 crc kubenswrapper[4575]: I1004 05:12:48.940691 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-ssh-key-openstack-edpm-ipam\") pod \"eff0e85f-146c-4e9f-8109-853df39fb0cd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " Oct 04 05:12:48 crc kubenswrapper[4575]: I1004 05:12:48.940781 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-inventory-0\") pod \"eff0e85f-146c-4e9f-8109-853df39fb0cd\" (UID: \"eff0e85f-146c-4e9f-8109-853df39fb0cd\") " Oct 04 05:12:48 crc kubenswrapper[4575]: I1004 05:12:48.950066 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff0e85f-146c-4e9f-8109-853df39fb0cd-kube-api-access-9gg49" (OuterVolumeSpecName: "kube-api-access-9gg49") pod "eff0e85f-146c-4e9f-8109-853df39fb0cd" (UID: "eff0e85f-146c-4e9f-8109-853df39fb0cd"). InnerVolumeSpecName "kube-api-access-9gg49". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:12:48 crc kubenswrapper[4575]: I1004 05:12:48.977568 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "eff0e85f-146c-4e9f-8109-853df39fb0cd" (UID: "eff0e85f-146c-4e9f-8109-853df39fb0cd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:12:48 crc kubenswrapper[4575]: I1004 05:12:48.981065 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "eff0e85f-146c-4e9f-8109-853df39fb0cd" (UID: "eff0e85f-146c-4e9f-8109-853df39fb0cd"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.042905 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gg49\" (UniqueName: \"kubernetes.io/projected/eff0e85f-146c-4e9f-8109-853df39fb0cd-kube-api-access-9gg49\") on node \"crc\" DevicePath \"\"" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.042944 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.042958 4575 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/eff0e85f-146c-4e9f-8109-853df39fb0cd-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.384369 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" event={"ID":"eff0e85f-146c-4e9f-8109-853df39fb0cd","Type":"ContainerDied","Data":"eadfc33dc4eed1f6ccd3a1875fafb77a5b3fcb267b38b5a9b379135f063ffd56"} Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.384668 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eadfc33dc4eed1f6ccd3a1875fafb77a5b3fcb267b38b5a9b379135f063ffd56" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.384616 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-j4phd" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.535650 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r"] Oct 04 05:12:49 crc kubenswrapper[4575]: E1004 05:12:49.536118 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff0e85f-146c-4e9f-8109-853df39fb0cd" containerName="ssh-known-hosts-edpm-deployment" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.536143 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff0e85f-146c-4e9f-8109-853df39fb0cd" containerName="ssh-known-hosts-edpm-deployment" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.536349 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff0e85f-146c-4e9f-8109-853df39fb0cd" containerName="ssh-known-hosts-edpm-deployment" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.537038 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.542413 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.542784 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.543059 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.554211 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.564813 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r"] Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.669515 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qcqj\" (UniqueName: \"kubernetes.io/projected/f7139366-9c39-49e0-a48f-63dd47e2f52f-kube-api-access-7qcqj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rjf7r\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.669715 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rjf7r\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.669786 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rjf7r\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.771871 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rjf7r\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.772168 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rjf7r\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.772322 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qcqj\" (UniqueName: \"kubernetes.io/projected/f7139366-9c39-49e0-a48f-63dd47e2f52f-kube-api-access-7qcqj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rjf7r\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.789499 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rjf7r\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.791307 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rjf7r\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.791744 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qcqj\" (UniqueName: \"kubernetes.io/projected/f7139366-9c39-49e0-a48f-63dd47e2f52f-kube-api-access-7qcqj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rjf7r\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:49 crc kubenswrapper[4575]: I1004 05:12:49.873003 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:12:50 crc kubenswrapper[4575]: I1004 05:12:50.310847 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:12:50 crc kubenswrapper[4575]: E1004 05:12:50.311390 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:12:50 crc kubenswrapper[4575]: I1004 05:12:50.428711 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r"] Oct 04 05:12:51 crc kubenswrapper[4575]: I1004 05:12:51.401710 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" event={"ID":"f7139366-9c39-49e0-a48f-63dd47e2f52f","Type":"ContainerStarted","Data":"79dc96f5a0eeb400c9b13dbedeabf8323c12ec2a28cd44af60b32bf8a5891940"} Oct 04 05:12:51 crc kubenswrapper[4575]: I1004 05:12:51.402265 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" event={"ID":"f7139366-9c39-49e0-a48f-63dd47e2f52f","Type":"ContainerStarted","Data":"0eeebf57341c8edcfaf3635dd14bf12fc4a8ae47e386388ee2bfe16c5a2dce41"} Oct 04 05:12:51 crc kubenswrapper[4575]: I1004 05:12:51.424006 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" podStartSLOduration=1.936090648 podStartE2EDuration="2.423986943s" podCreationTimestamp="2025-10-04 05:12:49 +0000 UTC" firstStartedPulling="2025-10-04 05:12:50.432915781 +0000 UTC m=+2321.761474595" lastFinishedPulling="2025-10-04 05:12:50.920812076 +0000 UTC m=+2322.249370890" observedRunningTime="2025-10-04 05:12:51.414246451 +0000 UTC m=+2322.742805265" watchObservedRunningTime="2025-10-04 05:12:51.423986943 +0000 UTC m=+2322.752545757" Oct 04 05:13:00 crc kubenswrapper[4575]: I1004 05:13:00.480014 4575 generic.go:334] "Generic (PLEG): container finished" podID="f7139366-9c39-49e0-a48f-63dd47e2f52f" containerID="79dc96f5a0eeb400c9b13dbedeabf8323c12ec2a28cd44af60b32bf8a5891940" exitCode=0 Oct 04 05:13:00 crc kubenswrapper[4575]: I1004 05:13:00.480106 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" event={"ID":"f7139366-9c39-49e0-a48f-63dd47e2f52f","Type":"ContainerDied","Data":"79dc96f5a0eeb400c9b13dbedeabf8323c12ec2a28cd44af60b32bf8a5891940"} Oct 04 05:13:01 crc kubenswrapper[4575]: I1004 05:13:01.905465 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.014957 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-ssh-key\") pod \"f7139366-9c39-49e0-a48f-63dd47e2f52f\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.015072 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qcqj\" (UniqueName: \"kubernetes.io/projected/f7139366-9c39-49e0-a48f-63dd47e2f52f-kube-api-access-7qcqj\") pod \"f7139366-9c39-49e0-a48f-63dd47e2f52f\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.015214 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-inventory\") pod \"f7139366-9c39-49e0-a48f-63dd47e2f52f\" (UID: \"f7139366-9c39-49e0-a48f-63dd47e2f52f\") " Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.033066 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7139366-9c39-49e0-a48f-63dd47e2f52f-kube-api-access-7qcqj" (OuterVolumeSpecName: "kube-api-access-7qcqj") pod "f7139366-9c39-49e0-a48f-63dd47e2f52f" (UID: "f7139366-9c39-49e0-a48f-63dd47e2f52f"). InnerVolumeSpecName "kube-api-access-7qcqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.057566 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f7139366-9c39-49e0-a48f-63dd47e2f52f" (UID: "f7139366-9c39-49e0-a48f-63dd47e2f52f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.057771 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-inventory" (OuterVolumeSpecName: "inventory") pod "f7139366-9c39-49e0-a48f-63dd47e2f52f" (UID: "f7139366-9c39-49e0-a48f-63dd47e2f52f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.117353 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.117397 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7139366-9c39-49e0-a48f-63dd47e2f52f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.117407 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qcqj\" (UniqueName: \"kubernetes.io/projected/f7139366-9c39-49e0-a48f-63dd47e2f52f-kube-api-access-7qcqj\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.497739 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" event={"ID":"f7139366-9c39-49e0-a48f-63dd47e2f52f","Type":"ContainerDied","Data":"0eeebf57341c8edcfaf3635dd14bf12fc4a8ae47e386388ee2bfe16c5a2dce41"} Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.497778 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0eeebf57341c8edcfaf3635dd14bf12fc4a8ae47e386388ee2bfe16c5a2dce41" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.497831 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rjf7r" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.590095 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd"] Oct 04 05:13:02 crc kubenswrapper[4575]: E1004 05:13:02.590547 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7139366-9c39-49e0-a48f-63dd47e2f52f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.590577 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7139366-9c39-49e0-a48f-63dd47e2f52f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.590804 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7139366-9c39-49e0-a48f-63dd47e2f52f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.594518 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.597708 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.597742 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.599620 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.599833 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.603080 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd"] Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.627534 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.627645 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b95sl\" (UniqueName: \"kubernetes.io/projected/32a6e686-a532-4a82-9132-4c466a1dc721-kube-api-access-b95sl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.627846 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.729873 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.729963 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.729998 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b95sl\" (UniqueName: \"kubernetes.io/projected/32a6e686-a532-4a82-9132-4c466a1dc721-kube-api-access-b95sl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.735365 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.738235 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.758175 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b95sl\" (UniqueName: \"kubernetes.io/projected/32a6e686-a532-4a82-9132-4c466a1dc721-kube-api-access-b95sl\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:02 crc kubenswrapper[4575]: I1004 05:13:02.920719 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:03 crc kubenswrapper[4575]: I1004 05:13:03.309982 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:13:03 crc kubenswrapper[4575]: E1004 05:13:03.310798 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:13:03 crc kubenswrapper[4575]: I1004 05:13:03.473049 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd"] Oct 04 05:13:03 crc kubenswrapper[4575]: I1004 05:13:03.514822 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" event={"ID":"32a6e686-a532-4a82-9132-4c466a1dc721","Type":"ContainerStarted","Data":"de4f0b662742d77cd969c5dfcf95c653975bc6ff90fab5eb798b5ff269b777c0"} Oct 04 05:13:05 crc kubenswrapper[4575]: I1004 05:13:05.570931 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" event={"ID":"32a6e686-a532-4a82-9132-4c466a1dc721","Type":"ContainerStarted","Data":"7897db71db8ab95fae41373d8af7db666ee37bf1df8f6b27225bc5ffacfae642"} Oct 04 05:13:05 crc kubenswrapper[4575]: I1004 05:13:05.599775 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" podStartSLOduration=2.914521369 podStartE2EDuration="3.599753697s" podCreationTimestamp="2025-10-04 05:13:02 +0000 UTC" firstStartedPulling="2025-10-04 05:13:03.487499755 +0000 UTC m=+2334.816058569" lastFinishedPulling="2025-10-04 05:13:04.172732083 +0000 UTC m=+2335.501290897" observedRunningTime="2025-10-04 05:13:05.594728042 +0000 UTC m=+2336.923286866" watchObservedRunningTime="2025-10-04 05:13:05.599753697 +0000 UTC m=+2336.928312511" Oct 04 05:13:14 crc kubenswrapper[4575]: I1004 05:13:14.644043 4575 generic.go:334] "Generic (PLEG): container finished" podID="32a6e686-a532-4a82-9132-4c466a1dc721" containerID="7897db71db8ab95fae41373d8af7db666ee37bf1df8f6b27225bc5ffacfae642" exitCode=0 Oct 04 05:13:14 crc kubenswrapper[4575]: I1004 05:13:14.644195 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" event={"ID":"32a6e686-a532-4a82-9132-4c466a1dc721","Type":"ContainerDied","Data":"7897db71db8ab95fae41373d8af7db666ee37bf1df8f6b27225bc5ffacfae642"} Oct 04 05:13:15 crc kubenswrapper[4575]: I1004 05:13:15.311575 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:13:15 crc kubenswrapper[4575]: E1004 05:13:15.312084 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.052490 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.122074 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-ssh-key\") pod \"32a6e686-a532-4a82-9132-4c466a1dc721\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.122191 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b95sl\" (UniqueName: \"kubernetes.io/projected/32a6e686-a532-4a82-9132-4c466a1dc721-kube-api-access-b95sl\") pod \"32a6e686-a532-4a82-9132-4c466a1dc721\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.122298 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-inventory\") pod \"32a6e686-a532-4a82-9132-4c466a1dc721\" (UID: \"32a6e686-a532-4a82-9132-4c466a1dc721\") " Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.127923 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32a6e686-a532-4a82-9132-4c466a1dc721-kube-api-access-b95sl" (OuterVolumeSpecName: "kube-api-access-b95sl") pod "32a6e686-a532-4a82-9132-4c466a1dc721" (UID: "32a6e686-a532-4a82-9132-4c466a1dc721"). InnerVolumeSpecName "kube-api-access-b95sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.150638 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-inventory" (OuterVolumeSpecName: "inventory") pod "32a6e686-a532-4a82-9132-4c466a1dc721" (UID: "32a6e686-a532-4a82-9132-4c466a1dc721"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.154971 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "32a6e686-a532-4a82-9132-4c466a1dc721" (UID: "32a6e686-a532-4a82-9132-4c466a1dc721"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.227078 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.227111 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b95sl\" (UniqueName: \"kubernetes.io/projected/32a6e686-a532-4a82-9132-4c466a1dc721-kube-api-access-b95sl\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.227125 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32a6e686-a532-4a82-9132-4c466a1dc721-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.666795 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" event={"ID":"32a6e686-a532-4a82-9132-4c466a1dc721","Type":"ContainerDied","Data":"de4f0b662742d77cd969c5dfcf95c653975bc6ff90fab5eb798b5ff269b777c0"} Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.666834 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de4f0b662742d77cd969c5dfcf95c653975bc6ff90fab5eb798b5ff269b777c0" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.666845 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.749248 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8"] Oct 04 05:13:16 crc kubenswrapper[4575]: E1004 05:13:16.749683 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a6e686-a532-4a82-9132-4c466a1dc721" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.749702 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a6e686-a532-4a82-9132-4c466a1dc721" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.749887 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="32a6e686-a532-4a82-9132-4c466a1dc721" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.750523 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.753886 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.754060 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.754192 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.754331 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.754697 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.754931 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.755075 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.758761 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.765687 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8"] Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.840251 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.840321 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.840358 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.840385 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.840409 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.840525 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.840559 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.840863 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.840941 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.841052 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.841103 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.841140 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.841226 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8t8r\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-kube-api-access-c8t8r\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.841354 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.943018 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.943318 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.943500 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.944048 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.944188 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.944331 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.944510 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.944896 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8t8r\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-kube-api-access-c8t8r\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.944999 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.945155 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.945244 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.945314 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.945399 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.945508 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.948422 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.948438 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.949602 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.953226 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.955249 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.955733 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.956408 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.956435 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.957954 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.962848 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.963894 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.967204 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8t8r\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-kube-api-access-c8t8r\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.967231 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:16 crc kubenswrapper[4575]: I1004 05:13:16.975369 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-x28j8\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:17 crc kubenswrapper[4575]: I1004 05:13:17.073114 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:17 crc kubenswrapper[4575]: I1004 05:13:17.432520 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8"] Oct 04 05:13:17 crc kubenswrapper[4575]: I1004 05:13:17.675783 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" event={"ID":"24f35162-c12e-4f8b-bce9-798fbf9b38bc","Type":"ContainerStarted","Data":"1cd1fff912d392e602560f4ad3317717af95ad74a51be8cd79b7c93c83667c1d"} Oct 04 05:13:18 crc kubenswrapper[4575]: I1004 05:13:18.687572 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" event={"ID":"24f35162-c12e-4f8b-bce9-798fbf9b38bc","Type":"ContainerStarted","Data":"8066d3a990cffcf594d60f8bc133faf7e985c708294ee446548412ea1bec5a99"} Oct 04 05:13:18 crc kubenswrapper[4575]: I1004 05:13:18.710043 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" podStartSLOduration=2.006094144 podStartE2EDuration="2.710017773s" podCreationTimestamp="2025-10-04 05:13:16 +0000 UTC" firstStartedPulling="2025-10-04 05:13:17.439465339 +0000 UTC m=+2348.768024153" lastFinishedPulling="2025-10-04 05:13:18.143388968 +0000 UTC m=+2349.471947782" observedRunningTime="2025-10-04 05:13:18.709268261 +0000 UTC m=+2350.037827085" watchObservedRunningTime="2025-10-04 05:13:18.710017773 +0000 UTC m=+2350.038576587" Oct 04 05:13:30 crc kubenswrapper[4575]: I1004 05:13:30.310205 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:13:30 crc kubenswrapper[4575]: E1004 05:13:30.310993 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:13:42 crc kubenswrapper[4575]: I1004 05:13:42.310566 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:13:42 crc kubenswrapper[4575]: E1004 05:13:42.311748 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:13:56 crc kubenswrapper[4575]: I1004 05:13:56.310602 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:13:56 crc kubenswrapper[4575]: E1004 05:13:56.311267 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:13:58 crc kubenswrapper[4575]: I1004 05:13:58.025374 4575 generic.go:334] "Generic (PLEG): container finished" podID="24f35162-c12e-4f8b-bce9-798fbf9b38bc" containerID="8066d3a990cffcf594d60f8bc133faf7e985c708294ee446548412ea1bec5a99" exitCode=0 Oct 04 05:13:58 crc kubenswrapper[4575]: I1004 05:13:58.025475 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" event={"ID":"24f35162-c12e-4f8b-bce9-798fbf9b38bc","Type":"ContainerDied","Data":"8066d3a990cffcf594d60f8bc133faf7e985c708294ee446548412ea1bec5a99"} Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.468826 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602122 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-inventory\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602212 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8t8r\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-kube-api-access-c8t8r\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602242 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-repo-setup-combined-ca-bundle\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602258 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-neutron-metadata-combined-ca-bundle\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602316 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ssh-key\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602345 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602376 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602413 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-nova-combined-ca-bundle\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602446 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-libvirt-combined-ca-bundle\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602552 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602612 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ovn-combined-ca-bundle\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602636 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-telemetry-combined-ca-bundle\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602685 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.602712 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-bootstrap-combined-ca-bundle\") pod \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\" (UID: \"24f35162-c12e-4f8b-bce9-798fbf9b38bc\") " Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.613442 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-kube-api-access-c8t8r" (OuterVolumeSpecName: "kube-api-access-c8t8r") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "kube-api-access-c8t8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.614894 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.615369 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.615697 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.616036 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.616118 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.616126 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.618609 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.621494 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.625974 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.626006 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.626196 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.643932 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.657479 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-inventory" (OuterVolumeSpecName: "inventory") pod "24f35162-c12e-4f8b-bce9-798fbf9b38bc" (UID: "24f35162-c12e-4f8b-bce9-798fbf9b38bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706194 4575 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706248 4575 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706260 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706275 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8t8r\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-kube-api-access-c8t8r\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706705 4575 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706716 4575 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706728 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706737 4575 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706767 4575 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706778 4575 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706787 4575 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706797 4575 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/24f35162-c12e-4f8b-bce9-798fbf9b38bc-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706807 4575 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:59 crc kubenswrapper[4575]: I1004 05:13:59.706817 4575 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f35162-c12e-4f8b-bce9-798fbf9b38bc-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.045746 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" event={"ID":"24f35162-c12e-4f8b-bce9-798fbf9b38bc","Type":"ContainerDied","Data":"1cd1fff912d392e602560f4ad3317717af95ad74a51be8cd79b7c93c83667c1d"} Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.046050 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cd1fff912d392e602560f4ad3317717af95ad74a51be8cd79b7c93c83667c1d" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.045801 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-x28j8" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.150031 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97"] Oct 04 05:14:00 crc kubenswrapper[4575]: E1004 05:14:00.150775 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f35162-c12e-4f8b-bce9-798fbf9b38bc" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.150800 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f35162-c12e-4f8b-bce9-798fbf9b38bc" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.151038 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f35162-c12e-4f8b-bce9-798fbf9b38bc" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.151780 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.153689 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.158121 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.158615 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.158946 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.161254 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.162298 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97"] Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.216844 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.216905 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.217007 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.217029 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h4bk\" (UniqueName: \"kubernetes.io/projected/e6092d98-ab7f-4313-86ad-783fa9f625dd-kube-api-access-7h4bk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.217076 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.321640 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.321768 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.322031 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.322068 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h4bk\" (UniqueName: \"kubernetes.io/projected/e6092d98-ab7f-4313-86ad-783fa9f625dd-kube-api-access-7h4bk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.322177 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.322872 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.325395 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.327457 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.328963 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.340244 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h4bk\" (UniqueName: \"kubernetes.io/projected/e6092d98-ab7f-4313-86ad-783fa9f625dd-kube-api-access-7h4bk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hdg97\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.472226 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:14:00 crc kubenswrapper[4575]: I1004 05:14:00.984865 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97"] Oct 04 05:14:00 crc kubenswrapper[4575]: W1004 05:14:00.989097 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6092d98_ab7f_4313_86ad_783fa9f625dd.slice/crio-110daa5ebcc27939afb4f5f1ec7461386693a6cc1c4adb7cfc2b612bcdc9e192 WatchSource:0}: Error finding container 110daa5ebcc27939afb4f5f1ec7461386693a6cc1c4adb7cfc2b612bcdc9e192: Status 404 returned error can't find the container with id 110daa5ebcc27939afb4f5f1ec7461386693a6cc1c4adb7cfc2b612bcdc9e192 Oct 04 05:14:01 crc kubenswrapper[4575]: I1004 05:14:01.065146 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" event={"ID":"e6092d98-ab7f-4313-86ad-783fa9f625dd","Type":"ContainerStarted","Data":"110daa5ebcc27939afb4f5f1ec7461386693a6cc1c4adb7cfc2b612bcdc9e192"} Oct 04 05:14:03 crc kubenswrapper[4575]: I1004 05:14:03.086676 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" event={"ID":"e6092d98-ab7f-4313-86ad-783fa9f625dd","Type":"ContainerStarted","Data":"a4229605a3263a2a2ff4871e55f7e89408d2fec8496d9aa65c0bac2f2ee8636c"} Oct 04 05:14:03 crc kubenswrapper[4575]: I1004 05:14:03.110922 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" podStartSLOduration=2.296327342 podStartE2EDuration="3.110902025s" podCreationTimestamp="2025-10-04 05:14:00 +0000 UTC" firstStartedPulling="2025-10-04 05:14:00.999797776 +0000 UTC m=+2392.328356590" lastFinishedPulling="2025-10-04 05:14:01.814372459 +0000 UTC m=+2393.142931273" observedRunningTime="2025-10-04 05:14:03.103149151 +0000 UTC m=+2394.431707965" watchObservedRunningTime="2025-10-04 05:14:03.110902025 +0000 UTC m=+2394.439460839" Oct 04 05:14:10 crc kubenswrapper[4575]: I1004 05:14:10.310908 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:14:10 crc kubenswrapper[4575]: E1004 05:14:10.311575 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:14:21 crc kubenswrapper[4575]: I1004 05:14:21.310922 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:14:21 crc kubenswrapper[4575]: E1004 05:14:21.312988 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:14:33 crc kubenswrapper[4575]: I1004 05:14:33.309692 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:14:33 crc kubenswrapper[4575]: E1004 05:14:33.310523 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:14:44 crc kubenswrapper[4575]: I1004 05:14:44.310485 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:14:44 crc kubenswrapper[4575]: E1004 05:14:44.311615 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:14:58 crc kubenswrapper[4575]: I1004 05:14:58.310937 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:14:58 crc kubenswrapper[4575]: E1004 05:14:58.311850 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.154693 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9"] Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.157218 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.159873 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.161995 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.170569 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9"] Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.342173 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zxqs\" (UniqueName: \"kubernetes.io/projected/64a0174c-ba82-4d55-b84c-fe28237cd599-kube-api-access-7zxqs\") pod \"collect-profiles-29325915-mnjm9\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.342253 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64a0174c-ba82-4d55-b84c-fe28237cd599-secret-volume\") pod \"collect-profiles-29325915-mnjm9\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.342389 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64a0174c-ba82-4d55-b84c-fe28237cd599-config-volume\") pod \"collect-profiles-29325915-mnjm9\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.444688 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zxqs\" (UniqueName: \"kubernetes.io/projected/64a0174c-ba82-4d55-b84c-fe28237cd599-kube-api-access-7zxqs\") pod \"collect-profiles-29325915-mnjm9\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.444828 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64a0174c-ba82-4d55-b84c-fe28237cd599-secret-volume\") pod \"collect-profiles-29325915-mnjm9\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.444928 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64a0174c-ba82-4d55-b84c-fe28237cd599-config-volume\") pod \"collect-profiles-29325915-mnjm9\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.446645 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64a0174c-ba82-4d55-b84c-fe28237cd599-config-volume\") pod \"collect-profiles-29325915-mnjm9\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.458822 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64a0174c-ba82-4d55-b84c-fe28237cd599-secret-volume\") pod \"collect-profiles-29325915-mnjm9\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.467804 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zxqs\" (UniqueName: \"kubernetes.io/projected/64a0174c-ba82-4d55-b84c-fe28237cd599-kube-api-access-7zxqs\") pod \"collect-profiles-29325915-mnjm9\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.497293 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:00 crc kubenswrapper[4575]: I1004 05:15:00.981292 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9"] Oct 04 05:15:01 crc kubenswrapper[4575]: I1004 05:15:01.642547 4575 generic.go:334] "Generic (PLEG): container finished" podID="64a0174c-ba82-4d55-b84c-fe28237cd599" containerID="82d2995568e20102a359e4524745e12f70b27edec15f20faf785afb4dabfeae9" exitCode=0 Oct 04 05:15:01 crc kubenswrapper[4575]: I1004 05:15:01.642621 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" event={"ID":"64a0174c-ba82-4d55-b84c-fe28237cd599","Type":"ContainerDied","Data":"82d2995568e20102a359e4524745e12f70b27edec15f20faf785afb4dabfeae9"} Oct 04 05:15:01 crc kubenswrapper[4575]: I1004 05:15:01.642932 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" event={"ID":"64a0174c-ba82-4d55-b84c-fe28237cd599","Type":"ContainerStarted","Data":"a4833fb2a5c934946671b83b2f8ad6e9cb47361be5b9e476b94a3f354a16d3ed"} Oct 04 05:15:02 crc kubenswrapper[4575]: I1004 05:15:02.968200 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.098388 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64a0174c-ba82-4d55-b84c-fe28237cd599-config-volume\") pod \"64a0174c-ba82-4d55-b84c-fe28237cd599\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.099283 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64a0174c-ba82-4d55-b84c-fe28237cd599-config-volume" (OuterVolumeSpecName: "config-volume") pod "64a0174c-ba82-4d55-b84c-fe28237cd599" (UID: "64a0174c-ba82-4d55-b84c-fe28237cd599"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.099448 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64a0174c-ba82-4d55-b84c-fe28237cd599-secret-volume\") pod \"64a0174c-ba82-4d55-b84c-fe28237cd599\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.099531 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zxqs\" (UniqueName: \"kubernetes.io/projected/64a0174c-ba82-4d55-b84c-fe28237cd599-kube-api-access-7zxqs\") pod \"64a0174c-ba82-4d55-b84c-fe28237cd599\" (UID: \"64a0174c-ba82-4d55-b84c-fe28237cd599\") " Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.100014 4575 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64a0174c-ba82-4d55-b84c-fe28237cd599-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.104908 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64a0174c-ba82-4d55-b84c-fe28237cd599-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "64a0174c-ba82-4d55-b84c-fe28237cd599" (UID: "64a0174c-ba82-4d55-b84c-fe28237cd599"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.107537 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64a0174c-ba82-4d55-b84c-fe28237cd599-kube-api-access-7zxqs" (OuterVolumeSpecName: "kube-api-access-7zxqs") pod "64a0174c-ba82-4d55-b84c-fe28237cd599" (UID: "64a0174c-ba82-4d55-b84c-fe28237cd599"). InnerVolumeSpecName "kube-api-access-7zxqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.201433 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zxqs\" (UniqueName: \"kubernetes.io/projected/64a0174c-ba82-4d55-b84c-fe28237cd599-kube-api-access-7zxqs\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.201478 4575 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64a0174c-ba82-4d55-b84c-fe28237cd599-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.662464 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" event={"ID":"64a0174c-ba82-4d55-b84c-fe28237cd599","Type":"ContainerDied","Data":"a4833fb2a5c934946671b83b2f8ad6e9cb47361be5b9e476b94a3f354a16d3ed"} Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.662513 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9" Oct 04 05:15:03 crc kubenswrapper[4575]: I1004 05:15:03.662514 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4833fb2a5c934946671b83b2f8ad6e9cb47361be5b9e476b94a3f354a16d3ed" Oct 04 05:15:04 crc kubenswrapper[4575]: I1004 05:15:04.067724 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g"] Oct 04 05:15:04 crc kubenswrapper[4575]: I1004 05:15:04.076339 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-d7k6g"] Oct 04 05:15:05 crc kubenswrapper[4575]: I1004 05:15:05.324908 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4caea03c-cd82-4798-b8b1-7007cc09079a" path="/var/lib/kubelet/pods/4caea03c-cd82-4798-b8b1-7007cc09079a/volumes" Oct 04 05:15:09 crc kubenswrapper[4575]: I1004 05:15:09.328266 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:15:09 crc kubenswrapper[4575]: E1004 05:15:09.329875 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:15:10 crc kubenswrapper[4575]: I1004 05:15:10.734460 4575 generic.go:334] "Generic (PLEG): container finished" podID="e6092d98-ab7f-4313-86ad-783fa9f625dd" containerID="a4229605a3263a2a2ff4871e55f7e89408d2fec8496d9aa65c0bac2f2ee8636c" exitCode=0 Oct 04 05:15:10 crc kubenswrapper[4575]: I1004 05:15:10.734636 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" event={"ID":"e6092d98-ab7f-4313-86ad-783fa9f625dd","Type":"ContainerDied","Data":"a4229605a3263a2a2ff4871e55f7e89408d2fec8496d9aa65c0bac2f2ee8636c"} Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.198424 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.210254 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h4bk\" (UniqueName: \"kubernetes.io/projected/e6092d98-ab7f-4313-86ad-783fa9f625dd-kube-api-access-7h4bk\") pod \"e6092d98-ab7f-4313-86ad-783fa9f625dd\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.210342 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovncontroller-config-0\") pod \"e6092d98-ab7f-4313-86ad-783fa9f625dd\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.210366 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-inventory\") pod \"e6092d98-ab7f-4313-86ad-783fa9f625dd\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.210441 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovn-combined-ca-bundle\") pod \"e6092d98-ab7f-4313-86ad-783fa9f625dd\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.210467 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ssh-key\") pod \"e6092d98-ab7f-4313-86ad-783fa9f625dd\" (UID: \"e6092d98-ab7f-4313-86ad-783fa9f625dd\") " Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.249173 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e6092d98-ab7f-4313-86ad-783fa9f625dd" (UID: "e6092d98-ab7f-4313-86ad-783fa9f625dd"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.258967 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6092d98-ab7f-4313-86ad-783fa9f625dd-kube-api-access-7h4bk" (OuterVolumeSpecName: "kube-api-access-7h4bk") pod "e6092d98-ab7f-4313-86ad-783fa9f625dd" (UID: "e6092d98-ab7f-4313-86ad-783fa9f625dd"). InnerVolumeSpecName "kube-api-access-7h4bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.266942 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e6092d98-ab7f-4313-86ad-783fa9f625dd" (UID: "e6092d98-ab7f-4313-86ad-783fa9f625dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.283629 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "e6092d98-ab7f-4313-86ad-783fa9f625dd" (UID: "e6092d98-ab7f-4313-86ad-783fa9f625dd"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.302216 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-inventory" (OuterVolumeSpecName: "inventory") pod "e6092d98-ab7f-4313-86ad-783fa9f625dd" (UID: "e6092d98-ab7f-4313-86ad-783fa9f625dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.313516 4575 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.313905 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.313915 4575 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.313924 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6092d98-ab7f-4313-86ad-783fa9f625dd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.313932 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h4bk\" (UniqueName: \"kubernetes.io/projected/e6092d98-ab7f-4313-86ad-783fa9f625dd-kube-api-access-7h4bk\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.759820 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" event={"ID":"e6092d98-ab7f-4313-86ad-783fa9f625dd","Type":"ContainerDied","Data":"110daa5ebcc27939afb4f5f1ec7461386693a6cc1c4adb7cfc2b612bcdc9e192"} Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.759865 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="110daa5ebcc27939afb4f5f1ec7461386693a6cc1c4adb7cfc2b612bcdc9e192" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.759931 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hdg97" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.843833 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw"] Oct 04 05:15:12 crc kubenswrapper[4575]: E1004 05:15:12.844274 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a0174c-ba82-4d55-b84c-fe28237cd599" containerName="collect-profiles" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.844295 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a0174c-ba82-4d55-b84c-fe28237cd599" containerName="collect-profiles" Oct 04 05:15:12 crc kubenswrapper[4575]: E1004 05:15:12.844318 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6092d98-ab7f-4313-86ad-783fa9f625dd" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.844327 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6092d98-ab7f-4313-86ad-783fa9f625dd" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.844558 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6092d98-ab7f-4313-86ad-783fa9f625dd" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.844614 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="64a0174c-ba82-4d55-b84c-fe28237cd599" containerName="collect-profiles" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.845380 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.848485 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.849023 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.849291 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.849572 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.851675 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.852510 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.863121 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw"] Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.925271 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.925454 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.925510 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.925558 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.925603 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:12 crc kubenswrapper[4575]: I1004 05:15:12.925703 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g52bc\" (UniqueName: \"kubernetes.io/projected/6618ef15-c06c-40a9-85a0-e071b9e13e4f-kube-api-access-g52bc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.027853 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g52bc\" (UniqueName: \"kubernetes.io/projected/6618ef15-c06c-40a9-85a0-e071b9e13e4f-kube-api-access-g52bc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.028029 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.028129 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.028178 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.028219 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.028254 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.035179 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.035547 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.035977 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.036367 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.036774 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.046996 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g52bc\" (UniqueName: \"kubernetes.io/projected/6618ef15-c06c-40a9-85a0-e071b9e13e4f-kube-api-access-g52bc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.165390 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.710445 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw"] Oct 04 05:15:13 crc kubenswrapper[4575]: I1004 05:15:13.805166 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" event={"ID":"6618ef15-c06c-40a9-85a0-e071b9e13e4f","Type":"ContainerStarted","Data":"eed2990d02347d619414ff147684b3786af94851c107532ef7b609718d42eea2"} Oct 04 05:15:14 crc kubenswrapper[4575]: I1004 05:15:14.815874 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" event={"ID":"6618ef15-c06c-40a9-85a0-e071b9e13e4f","Type":"ContainerStarted","Data":"03d54f6bb7530e2773eeabf22ae513e30d189c260a68ca579d3dac500d668d0a"} Oct 04 05:15:14 crc kubenswrapper[4575]: I1004 05:15:14.852725 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" podStartSLOduration=2.068118636 podStartE2EDuration="2.852702066s" podCreationTimestamp="2025-10-04 05:15:12 +0000 UTC" firstStartedPulling="2025-10-04 05:15:13.716184226 +0000 UTC m=+2465.044743040" lastFinishedPulling="2025-10-04 05:15:14.500767646 +0000 UTC m=+2465.829326470" observedRunningTime="2025-10-04 05:15:14.837943159 +0000 UTC m=+2466.166501973" watchObservedRunningTime="2025-10-04 05:15:14.852702066 +0000 UTC m=+2466.181260890" Oct 04 05:15:20 crc kubenswrapper[4575]: I1004 05:15:20.310258 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:15:20 crc kubenswrapper[4575]: E1004 05:15:20.311435 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:15:30 crc kubenswrapper[4575]: I1004 05:15:30.717233 4575 scope.go:117] "RemoveContainer" containerID="304fb6501316231bf2fa86a48b13d1c20f8edf18912b5bae16f7824e5a3067df" Oct 04 05:15:32 crc kubenswrapper[4575]: I1004 05:15:32.310975 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:15:32 crc kubenswrapper[4575]: E1004 05:15:32.311804 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:15:43 crc kubenswrapper[4575]: I1004 05:15:43.310886 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:15:43 crc kubenswrapper[4575]: E1004 05:15:43.311779 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:15:56 crc kubenswrapper[4575]: I1004 05:15:56.309496 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:15:56 crc kubenswrapper[4575]: E1004 05:15:56.310152 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:16:07 crc kubenswrapper[4575]: I1004 05:16:07.299285 4575 generic.go:334] "Generic (PLEG): container finished" podID="6618ef15-c06c-40a9-85a0-e071b9e13e4f" containerID="03d54f6bb7530e2773eeabf22ae513e30d189c260a68ca579d3dac500d668d0a" exitCode=0 Oct 04 05:16:07 crc kubenswrapper[4575]: I1004 05:16:07.299372 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" event={"ID":"6618ef15-c06c-40a9-85a0-e071b9e13e4f","Type":"ContainerDied","Data":"03d54f6bb7530e2773eeabf22ae513e30d189c260a68ca579d3dac500d668d0a"} Oct 04 05:16:07 crc kubenswrapper[4575]: I1004 05:16:07.316744 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:16:07 crc kubenswrapper[4575]: E1004 05:16:07.317045 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.778232 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.905675 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-inventory\") pod \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.906125 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-metadata-combined-ca-bundle\") pod \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.906310 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g52bc\" (UniqueName: \"kubernetes.io/projected/6618ef15-c06c-40a9-85a0-e071b9e13e4f-kube-api-access-g52bc\") pod \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.906400 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-ssh-key\") pod \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.906546 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.906738 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-nova-metadata-neutron-config-0\") pod \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\" (UID: \"6618ef15-c06c-40a9-85a0-e071b9e13e4f\") " Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.915934 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6618ef15-c06c-40a9-85a0-e071b9e13e4f-kube-api-access-g52bc" (OuterVolumeSpecName: "kube-api-access-g52bc") pod "6618ef15-c06c-40a9-85a0-e071b9e13e4f" (UID: "6618ef15-c06c-40a9-85a0-e071b9e13e4f"). InnerVolumeSpecName "kube-api-access-g52bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.932392 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-inventory" (OuterVolumeSpecName: "inventory") pod "6618ef15-c06c-40a9-85a0-e071b9e13e4f" (UID: "6618ef15-c06c-40a9-85a0-e071b9e13e4f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.935867 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6618ef15-c06c-40a9-85a0-e071b9e13e4f" (UID: "6618ef15-c06c-40a9-85a0-e071b9e13e4f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.936818 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "6618ef15-c06c-40a9-85a0-e071b9e13e4f" (UID: "6618ef15-c06c-40a9-85a0-e071b9e13e4f"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.940674 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "6618ef15-c06c-40a9-85a0-e071b9e13e4f" (UID: "6618ef15-c06c-40a9-85a0-e071b9e13e4f"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:08 crc kubenswrapper[4575]: I1004 05:16:08.952136 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6618ef15-c06c-40a9-85a0-e071b9e13e4f" (UID: "6618ef15-c06c-40a9-85a0-e071b9e13e4f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.009366 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.009416 4575 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.009433 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g52bc\" (UniqueName: \"kubernetes.io/projected/6618ef15-c06c-40a9-85a0-e071b9e13e4f-kube-api-access-g52bc\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.009446 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.009455 4575 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.009485 4575 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6618ef15-c06c-40a9-85a0-e071b9e13e4f-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.333261 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" event={"ID":"6618ef15-c06c-40a9-85a0-e071b9e13e4f","Type":"ContainerDied","Data":"eed2990d02347d619414ff147684b3786af94851c107532ef7b609718d42eea2"} Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.333567 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eed2990d02347d619414ff147684b3786af94851c107532ef7b609718d42eea2" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.333343 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.424137 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg"] Oct 04 05:16:09 crc kubenswrapper[4575]: E1004 05:16:09.424555 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6618ef15-c06c-40a9-85a0-e071b9e13e4f" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.424571 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="6618ef15-c06c-40a9-85a0-e071b9e13e4f" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.424781 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="6618ef15-c06c-40a9-85a0-e071b9e13e4f" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.425387 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.427322 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.427840 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.428211 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.428301 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.428671 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.450916 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg"] Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.518856 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvblq\" (UniqueName: \"kubernetes.io/projected/d143e777-6b76-40a2-886b-a64a66c5d467-kube-api-access-cvblq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.518996 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.519031 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.519065 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.519084 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.620531 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.620902 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.620997 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.621107 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.621231 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvblq\" (UniqueName: \"kubernetes.io/projected/d143e777-6b76-40a2-886b-a64a66c5d467-kube-api-access-cvblq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.625184 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.625370 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.625877 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.631145 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.639193 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvblq\" (UniqueName: \"kubernetes.io/projected/d143e777-6b76-40a2-886b-a64a66c5d467-kube-api-access-cvblq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:09 crc kubenswrapper[4575]: I1004 05:16:09.744635 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:16:10 crc kubenswrapper[4575]: I1004 05:16:10.295139 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg"] Oct 04 05:16:10 crc kubenswrapper[4575]: I1004 05:16:10.345172 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" event={"ID":"d143e777-6b76-40a2-886b-a64a66c5d467","Type":"ContainerStarted","Data":"58297dd409a0e6103b227effdf701d2e8523bac06f51a2ddb9f03cc88b5262c3"} Oct 04 05:16:11 crc kubenswrapper[4575]: I1004 05:16:11.357140 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" event={"ID":"d143e777-6b76-40a2-886b-a64a66c5d467","Type":"ContainerStarted","Data":"def0eecae0e450dda1bcffce51385ce0956bf580fbc484a9406cb2255a49344d"} Oct 04 05:16:11 crc kubenswrapper[4575]: I1004 05:16:11.388008 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" podStartSLOduration=1.992501823 podStartE2EDuration="2.387987366s" podCreationTimestamp="2025-10-04 05:16:09 +0000 UTC" firstStartedPulling="2025-10-04 05:16:10.299805106 +0000 UTC m=+2521.628363920" lastFinishedPulling="2025-10-04 05:16:10.695290649 +0000 UTC m=+2522.023849463" observedRunningTime="2025-10-04 05:16:11.379465269 +0000 UTC m=+2522.708024163" watchObservedRunningTime="2025-10-04 05:16:11.387987366 +0000 UTC m=+2522.716546180" Oct 04 05:16:21 crc kubenswrapper[4575]: I1004 05:16:21.310615 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:16:21 crc kubenswrapper[4575]: E1004 05:16:21.311703 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:16:32 crc kubenswrapper[4575]: I1004 05:16:32.310433 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:16:32 crc kubenswrapper[4575]: E1004 05:16:32.311361 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:16:43 crc kubenswrapper[4575]: I1004 05:16:43.309873 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:16:43 crc kubenswrapper[4575]: E1004 05:16:43.310679 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:16:56 crc kubenswrapper[4575]: I1004 05:16:56.310180 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:16:56 crc kubenswrapper[4575]: E1004 05:16:56.311046 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:17:07 crc kubenswrapper[4575]: I1004 05:17:07.310672 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:17:07 crc kubenswrapper[4575]: E1004 05:17:07.311645 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:17:21 crc kubenswrapper[4575]: I1004 05:17:21.309873 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:17:21 crc kubenswrapper[4575]: E1004 05:17:21.310826 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:17:36 crc kubenswrapper[4575]: I1004 05:17:36.310296 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:17:36 crc kubenswrapper[4575]: E1004 05:17:36.311095 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:17:48 crc kubenswrapper[4575]: I1004 05:17:48.309817 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:17:49 crc kubenswrapper[4575]: I1004 05:17:49.285316 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"b77dbadefa0a5842cfdc96afce016aff88e8e24fb9b590a5ca3d5e7bcec7c7aa"} Oct 04 05:20:08 crc kubenswrapper[4575]: I1004 05:20:08.446783 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:20:08 crc kubenswrapper[4575]: I1004 05:20:08.447424 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:20:38 crc kubenswrapper[4575]: I1004 05:20:38.446898 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:20:38 crc kubenswrapper[4575]: I1004 05:20:38.448771 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.494860 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kxt9v"] Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.497678 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.530341 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kxt9v"] Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.637922 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz94x\" (UniqueName: \"kubernetes.io/projected/72c41444-3a71-4d1b-8754-59ed922e4bc5-kube-api-access-pz94x\") pod \"certified-operators-kxt9v\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.638103 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-utilities\") pod \"certified-operators-kxt9v\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.638124 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-catalog-content\") pod \"certified-operators-kxt9v\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.739567 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-utilities\") pod \"certified-operators-kxt9v\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.739646 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-catalog-content\") pod \"certified-operators-kxt9v\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.739725 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz94x\" (UniqueName: \"kubernetes.io/projected/72c41444-3a71-4d1b-8754-59ed922e4bc5-kube-api-access-pz94x\") pod \"certified-operators-kxt9v\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.740697 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-utilities\") pod \"certified-operators-kxt9v\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.740927 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-catalog-content\") pod \"certified-operators-kxt9v\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.767534 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz94x\" (UniqueName: \"kubernetes.io/projected/72c41444-3a71-4d1b-8754-59ed922e4bc5-kube-api-access-pz94x\") pod \"certified-operators-kxt9v\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:41 crc kubenswrapper[4575]: I1004 05:20:41.834196 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:42 crc kubenswrapper[4575]: I1004 05:20:42.424873 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kxt9v"] Oct 04 05:20:42 crc kubenswrapper[4575]: I1004 05:20:42.823652 4575 generic.go:334] "Generic (PLEG): container finished" podID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerID="73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6" exitCode=0 Oct 04 05:20:42 crc kubenswrapper[4575]: I1004 05:20:42.823886 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxt9v" event={"ID":"72c41444-3a71-4d1b-8754-59ed922e4bc5","Type":"ContainerDied","Data":"73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6"} Oct 04 05:20:42 crc kubenswrapper[4575]: I1004 05:20:42.823947 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxt9v" event={"ID":"72c41444-3a71-4d1b-8754-59ed922e4bc5","Type":"ContainerStarted","Data":"c68fe32bd7ba7922da88a330b9397afeba9feeae88bfc022d9ad55e82993791b"} Oct 04 05:20:42 crc kubenswrapper[4575]: I1004 05:20:42.826331 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:20:44 crc kubenswrapper[4575]: I1004 05:20:44.845833 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxt9v" event={"ID":"72c41444-3a71-4d1b-8754-59ed922e4bc5","Type":"ContainerStarted","Data":"895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e"} Oct 04 05:20:46 crc kubenswrapper[4575]: I1004 05:20:46.870867 4575 generic.go:334] "Generic (PLEG): container finished" podID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerID="895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e" exitCode=0 Oct 04 05:20:46 crc kubenswrapper[4575]: I1004 05:20:46.870977 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxt9v" event={"ID":"72c41444-3a71-4d1b-8754-59ed922e4bc5","Type":"ContainerDied","Data":"895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e"} Oct 04 05:20:47 crc kubenswrapper[4575]: I1004 05:20:47.882997 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxt9v" event={"ID":"72c41444-3a71-4d1b-8754-59ed922e4bc5","Type":"ContainerStarted","Data":"7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69"} Oct 04 05:20:47 crc kubenswrapper[4575]: I1004 05:20:47.910323 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kxt9v" podStartSLOduration=2.367148567 podStartE2EDuration="6.910296746s" podCreationTimestamp="2025-10-04 05:20:41 +0000 UTC" firstStartedPulling="2025-10-04 05:20:42.82603719 +0000 UTC m=+2794.154596004" lastFinishedPulling="2025-10-04 05:20:47.369185359 +0000 UTC m=+2798.697744183" observedRunningTime="2025-10-04 05:20:47.904386715 +0000 UTC m=+2799.232945549" watchObservedRunningTime="2025-10-04 05:20:47.910296746 +0000 UTC m=+2799.238855570" Oct 04 05:20:51 crc kubenswrapper[4575]: I1004 05:20:51.835868 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:51 crc kubenswrapper[4575]: I1004 05:20:51.836439 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:51 crc kubenswrapper[4575]: I1004 05:20:51.885396 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:20:54 crc kubenswrapper[4575]: I1004 05:20:54.957755 4575 generic.go:334] "Generic (PLEG): container finished" podID="d143e777-6b76-40a2-886b-a64a66c5d467" containerID="def0eecae0e450dda1bcffce51385ce0956bf580fbc484a9406cb2255a49344d" exitCode=0 Oct 04 05:20:54 crc kubenswrapper[4575]: I1004 05:20:54.957832 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" event={"ID":"d143e777-6b76-40a2-886b-a64a66c5d467","Type":"ContainerDied","Data":"def0eecae0e450dda1bcffce51385ce0956bf580fbc484a9406cb2255a49344d"} Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.354733 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.447158 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-inventory\") pod \"d143e777-6b76-40a2-886b-a64a66c5d467\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.447545 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvblq\" (UniqueName: \"kubernetes.io/projected/d143e777-6b76-40a2-886b-a64a66c5d467-kube-api-access-cvblq\") pod \"d143e777-6b76-40a2-886b-a64a66c5d467\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.447833 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-combined-ca-bundle\") pod \"d143e777-6b76-40a2-886b-a64a66c5d467\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.448086 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-secret-0\") pod \"d143e777-6b76-40a2-886b-a64a66c5d467\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.448211 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-ssh-key\") pod \"d143e777-6b76-40a2-886b-a64a66c5d467\" (UID: \"d143e777-6b76-40a2-886b-a64a66c5d467\") " Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.465882 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d143e777-6b76-40a2-886b-a64a66c5d467" (UID: "d143e777-6b76-40a2-886b-a64a66c5d467"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.466257 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d143e777-6b76-40a2-886b-a64a66c5d467-kube-api-access-cvblq" (OuterVolumeSpecName: "kube-api-access-cvblq") pod "d143e777-6b76-40a2-886b-a64a66c5d467" (UID: "d143e777-6b76-40a2-886b-a64a66c5d467"). InnerVolumeSpecName "kube-api-access-cvblq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.492846 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "d143e777-6b76-40a2-886b-a64a66c5d467" (UID: "d143e777-6b76-40a2-886b-a64a66c5d467"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.493148 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-inventory" (OuterVolumeSpecName: "inventory") pod "d143e777-6b76-40a2-886b-a64a66c5d467" (UID: "d143e777-6b76-40a2-886b-a64a66c5d467"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.495348 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d143e777-6b76-40a2-886b-a64a66c5d467" (UID: "d143e777-6b76-40a2-886b-a64a66c5d467"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.551723 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.551869 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.551988 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvblq\" (UniqueName: \"kubernetes.io/projected/d143e777-6b76-40a2-886b-a64a66c5d467-kube-api-access-cvblq\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.552080 4575 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.552173 4575 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d143e777-6b76-40a2-886b-a64a66c5d467-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.976694 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.976667 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg" event={"ID":"d143e777-6b76-40a2-886b-a64a66c5d467","Type":"ContainerDied","Data":"58297dd409a0e6103b227effdf701d2e8523bac06f51a2ddb9f03cc88b5262c3"} Oct 04 05:20:56 crc kubenswrapper[4575]: I1004 05:20:56.977130 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58297dd409a0e6103b227effdf701d2e8523bac06f51a2ddb9f03cc88b5262c3" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.077361 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442"] Oct 04 05:20:57 crc kubenswrapper[4575]: E1004 05:20:57.077896 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d143e777-6b76-40a2-886b-a64a66c5d467" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.077923 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="d143e777-6b76-40a2-886b-a64a66c5d467" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.078193 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="d143e777-6b76-40a2-886b-a64a66c5d467" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.079035 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.082941 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.083335 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.083782 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.083914 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.084323 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.086362 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.086740 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.097722 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442"] Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.164781 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.165429 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.165629 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.165756 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mftjs\" (UniqueName: \"kubernetes.io/projected/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-kube-api-access-mftjs\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.165900 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.166016 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.166304 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.166393 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.166859 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.269823 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.270050 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.270095 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mftjs\" (UniqueName: \"kubernetes.io/projected/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-kube-api-access-mftjs\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.270173 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.270204 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.270314 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.270344 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.270512 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.270575 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.271772 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.275383 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.275848 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.276288 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.277073 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.279616 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.282928 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.293772 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.293845 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mftjs\" (UniqueName: \"kubernetes.io/projected/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-kube-api-access-mftjs\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ft442\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.399018 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:20:57 crc kubenswrapper[4575]: I1004 05:20:57.981541 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442"] Oct 04 05:20:58 crc kubenswrapper[4575]: I1004 05:20:58.994741 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" event={"ID":"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17","Type":"ContainerStarted","Data":"84af6207453a50e6bb95146d5972c4862974b300fc75f6524c1187e6525f060e"} Oct 04 05:21:00 crc kubenswrapper[4575]: I1004 05:21:00.004182 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" event={"ID":"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17","Type":"ContainerStarted","Data":"b4a62906508c4f5bed04c43401e4f2c71f6644eab6768ad65387bc0e0da7d5e2"} Oct 04 05:21:00 crc kubenswrapper[4575]: I1004 05:21:00.026547 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" podStartSLOduration=2.275698479 podStartE2EDuration="3.026532144s" podCreationTimestamp="2025-10-04 05:20:57 +0000 UTC" firstStartedPulling="2025-10-04 05:20:57.986558517 +0000 UTC m=+2809.315117331" lastFinishedPulling="2025-10-04 05:20:58.737392182 +0000 UTC m=+2810.065950996" observedRunningTime="2025-10-04 05:21:00.020261562 +0000 UTC m=+2811.348820376" watchObservedRunningTime="2025-10-04 05:21:00.026532144 +0000 UTC m=+2811.355090958" Oct 04 05:21:01 crc kubenswrapper[4575]: I1004 05:21:01.899903 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.002789 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kxt9v"] Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.019310 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kxt9v" podUID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerName="registry-server" containerID="cri-o://7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69" gracePeriod=2 Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.477152 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.606120 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-utilities\") pod \"72c41444-3a71-4d1b-8754-59ed922e4bc5\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.606563 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-catalog-content\") pod \"72c41444-3a71-4d1b-8754-59ed922e4bc5\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.606679 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz94x\" (UniqueName: \"kubernetes.io/projected/72c41444-3a71-4d1b-8754-59ed922e4bc5-kube-api-access-pz94x\") pod \"72c41444-3a71-4d1b-8754-59ed922e4bc5\" (UID: \"72c41444-3a71-4d1b-8754-59ed922e4bc5\") " Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.608319 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-utilities" (OuterVolumeSpecName: "utilities") pod "72c41444-3a71-4d1b-8754-59ed922e4bc5" (UID: "72c41444-3a71-4d1b-8754-59ed922e4bc5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.618804 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72c41444-3a71-4d1b-8754-59ed922e4bc5-kube-api-access-pz94x" (OuterVolumeSpecName: "kube-api-access-pz94x") pod "72c41444-3a71-4d1b-8754-59ed922e4bc5" (UID: "72c41444-3a71-4d1b-8754-59ed922e4bc5"). InnerVolumeSpecName "kube-api-access-pz94x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.660940 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72c41444-3a71-4d1b-8754-59ed922e4bc5" (UID: "72c41444-3a71-4d1b-8754-59ed922e4bc5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.709330 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.709380 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c41444-3a71-4d1b-8754-59ed922e4bc5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:02 crc kubenswrapper[4575]: I1004 05:21:02.709397 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz94x\" (UniqueName: \"kubernetes.io/projected/72c41444-3a71-4d1b-8754-59ed922e4bc5-kube-api-access-pz94x\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.029431 4575 generic.go:334] "Generic (PLEG): container finished" podID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerID="7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69" exitCode=0 Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.029476 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxt9v" event={"ID":"72c41444-3a71-4d1b-8754-59ed922e4bc5","Type":"ContainerDied","Data":"7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69"} Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.029511 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxt9v" event={"ID":"72c41444-3a71-4d1b-8754-59ed922e4bc5","Type":"ContainerDied","Data":"c68fe32bd7ba7922da88a330b9397afeba9feeae88bfc022d9ad55e82993791b"} Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.029535 4575 scope.go:117] "RemoveContainer" containerID="7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.030766 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxt9v" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.063658 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kxt9v"] Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.065571 4575 scope.go:117] "RemoveContainer" containerID="895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.073473 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kxt9v"] Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.087603 4575 scope.go:117] "RemoveContainer" containerID="73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.124728 4575 scope.go:117] "RemoveContainer" containerID="7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69" Oct 04 05:21:03 crc kubenswrapper[4575]: E1004 05:21:03.125117 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69\": container with ID starting with 7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69 not found: ID does not exist" containerID="7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.125205 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69"} err="failed to get container status \"7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69\": rpc error: code = NotFound desc = could not find container \"7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69\": container with ID starting with 7da872e992f9b3570d5be560da510ee9925d22b8770e920900f50c4c8a84ad69 not found: ID does not exist" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.125290 4575 scope.go:117] "RemoveContainer" containerID="895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e" Oct 04 05:21:03 crc kubenswrapper[4575]: E1004 05:21:03.125548 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e\": container with ID starting with 895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e not found: ID does not exist" containerID="895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.125609 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e"} err="failed to get container status \"895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e\": rpc error: code = NotFound desc = could not find container \"895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e\": container with ID starting with 895004ab13c18705b841655dd2d046c86ead75a21a73bd0e358cf2520e86c79e not found: ID does not exist" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.125627 4575 scope.go:117] "RemoveContainer" containerID="73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6" Oct 04 05:21:03 crc kubenswrapper[4575]: E1004 05:21:03.125846 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6\": container with ID starting with 73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6 not found: ID does not exist" containerID="73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.125924 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6"} err="failed to get container status \"73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6\": rpc error: code = NotFound desc = could not find container \"73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6\": container with ID starting with 73b192e5d08fa5f8a261783bbed372a007cf7f29b323400ec09a2974e25436e6 not found: ID does not exist" Oct 04 05:21:03 crc kubenswrapper[4575]: I1004 05:21:03.321058 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72c41444-3a71-4d1b-8754-59ed922e4bc5" path="/var/lib/kubelet/pods/72c41444-3a71-4d1b-8754-59ed922e4bc5/volumes" Oct 04 05:21:08 crc kubenswrapper[4575]: I1004 05:21:08.446129 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:21:08 crc kubenswrapper[4575]: I1004 05:21:08.446712 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:21:08 crc kubenswrapper[4575]: I1004 05:21:08.446772 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:21:08 crc kubenswrapper[4575]: I1004 05:21:08.447564 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b77dbadefa0a5842cfdc96afce016aff88e8e24fb9b590a5ca3d5e7bcec7c7aa"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:21:08 crc kubenswrapper[4575]: I1004 05:21:08.447668 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://b77dbadefa0a5842cfdc96afce016aff88e8e24fb9b590a5ca3d5e7bcec7c7aa" gracePeriod=600 Oct 04 05:21:09 crc kubenswrapper[4575]: I1004 05:21:09.090569 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="b77dbadefa0a5842cfdc96afce016aff88e8e24fb9b590a5ca3d5e7bcec7c7aa" exitCode=0 Oct 04 05:21:09 crc kubenswrapper[4575]: I1004 05:21:09.090634 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"b77dbadefa0a5842cfdc96afce016aff88e8e24fb9b590a5ca3d5e7bcec7c7aa"} Oct 04 05:21:09 crc kubenswrapper[4575]: I1004 05:21:09.090988 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26"} Oct 04 05:21:09 crc kubenswrapper[4575]: I1004 05:21:09.091011 4575 scope.go:117] "RemoveContainer" containerID="597c295ebe31adbfe107eb5f683db190cf13a4d433d47093c9a8d4be5c00287e" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.368629 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7qzkh"] Oct 04 05:21:12 crc kubenswrapper[4575]: E1004 05:21:12.369779 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerName="registry-server" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.369799 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerName="registry-server" Oct 04 05:21:12 crc kubenswrapper[4575]: E1004 05:21:12.369845 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerName="extract-utilities" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.369854 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerName="extract-utilities" Oct 04 05:21:12 crc kubenswrapper[4575]: E1004 05:21:12.369873 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerName="extract-content" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.369880 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerName="extract-content" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.370127 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c41444-3a71-4d1b-8754-59ed922e4bc5" containerName="registry-server" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.382233 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7qzkh"] Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.382341 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.502379 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-catalog-content\") pod \"redhat-operators-7qzkh\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.502502 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-utilities\") pod \"redhat-operators-7qzkh\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.502662 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw5s4\" (UniqueName: \"kubernetes.io/projected/2dda26c0-0897-4f06-8278-bfb5551ad2f8-kube-api-access-tw5s4\") pod \"redhat-operators-7qzkh\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.604765 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw5s4\" (UniqueName: \"kubernetes.io/projected/2dda26c0-0897-4f06-8278-bfb5551ad2f8-kube-api-access-tw5s4\") pod \"redhat-operators-7qzkh\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.604839 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-catalog-content\") pod \"redhat-operators-7qzkh\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.604909 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-utilities\") pod \"redhat-operators-7qzkh\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.605460 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-utilities\") pod \"redhat-operators-7qzkh\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.605485 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-catalog-content\") pod \"redhat-operators-7qzkh\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.635492 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw5s4\" (UniqueName: \"kubernetes.io/projected/2dda26c0-0897-4f06-8278-bfb5551ad2f8-kube-api-access-tw5s4\") pod \"redhat-operators-7qzkh\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:12 crc kubenswrapper[4575]: I1004 05:21:12.706297 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:13 crc kubenswrapper[4575]: I1004 05:21:13.254745 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7qzkh"] Oct 04 05:21:14 crc kubenswrapper[4575]: I1004 05:21:14.146560 4575 generic.go:334] "Generic (PLEG): container finished" podID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerID="2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b" exitCode=0 Oct 04 05:21:14 crc kubenswrapper[4575]: I1004 05:21:14.146661 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qzkh" event={"ID":"2dda26c0-0897-4f06-8278-bfb5551ad2f8","Type":"ContainerDied","Data":"2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b"} Oct 04 05:21:14 crc kubenswrapper[4575]: I1004 05:21:14.147021 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qzkh" event={"ID":"2dda26c0-0897-4f06-8278-bfb5551ad2f8","Type":"ContainerStarted","Data":"f18dbf83105cac8c45380deff22c904a0c6e34f2c756c7b21ec747c58ce030d1"} Oct 04 05:21:16 crc kubenswrapper[4575]: I1004 05:21:16.167871 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qzkh" event={"ID":"2dda26c0-0897-4f06-8278-bfb5551ad2f8","Type":"ContainerStarted","Data":"5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3"} Oct 04 05:21:19 crc kubenswrapper[4575]: I1004 05:21:19.202245 4575 generic.go:334] "Generic (PLEG): container finished" podID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerID="5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3" exitCode=0 Oct 04 05:21:19 crc kubenswrapper[4575]: I1004 05:21:19.202342 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qzkh" event={"ID":"2dda26c0-0897-4f06-8278-bfb5551ad2f8","Type":"ContainerDied","Data":"5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3"} Oct 04 05:21:20 crc kubenswrapper[4575]: I1004 05:21:20.218661 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qzkh" event={"ID":"2dda26c0-0897-4f06-8278-bfb5551ad2f8","Type":"ContainerStarted","Data":"58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f"} Oct 04 05:21:20 crc kubenswrapper[4575]: I1004 05:21:20.255451 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7qzkh" podStartSLOduration=2.750086086 podStartE2EDuration="8.255418296s" podCreationTimestamp="2025-10-04 05:21:12 +0000 UTC" firstStartedPulling="2025-10-04 05:21:14.150849408 +0000 UTC m=+2825.479408222" lastFinishedPulling="2025-10-04 05:21:19.656181608 +0000 UTC m=+2830.984740432" observedRunningTime="2025-10-04 05:21:20.242849782 +0000 UTC m=+2831.571408626" watchObservedRunningTime="2025-10-04 05:21:20.255418296 +0000 UTC m=+2831.583977110" Oct 04 05:21:22 crc kubenswrapper[4575]: I1004 05:21:22.706919 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:22 crc kubenswrapper[4575]: I1004 05:21:22.707911 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:23 crc kubenswrapper[4575]: I1004 05:21:23.757359 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7qzkh" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerName="registry-server" probeResult="failure" output=< Oct 04 05:21:23 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:21:23 crc kubenswrapper[4575]: > Oct 04 05:21:32 crc kubenswrapper[4575]: I1004 05:21:32.761161 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:32 crc kubenswrapper[4575]: I1004 05:21:32.819682 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:32 crc kubenswrapper[4575]: I1004 05:21:32.997150 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7qzkh"] Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.352370 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7qzkh" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerName="registry-server" containerID="cri-o://58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f" gracePeriod=2 Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.794642 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.867766 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw5s4\" (UniqueName: \"kubernetes.io/projected/2dda26c0-0897-4f06-8278-bfb5551ad2f8-kube-api-access-tw5s4\") pod \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.867871 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-utilities\") pod \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.867923 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-catalog-content\") pod \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\" (UID: \"2dda26c0-0897-4f06-8278-bfb5551ad2f8\") " Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.868496 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-utilities" (OuterVolumeSpecName: "utilities") pod "2dda26c0-0897-4f06-8278-bfb5551ad2f8" (UID: "2dda26c0-0897-4f06-8278-bfb5551ad2f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.874665 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dda26c0-0897-4f06-8278-bfb5551ad2f8-kube-api-access-tw5s4" (OuterVolumeSpecName: "kube-api-access-tw5s4") pod "2dda26c0-0897-4f06-8278-bfb5551ad2f8" (UID: "2dda26c0-0897-4f06-8278-bfb5551ad2f8"). InnerVolumeSpecName "kube-api-access-tw5s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.952119 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dda26c0-0897-4f06-8278-bfb5551ad2f8" (UID: "2dda26c0-0897-4f06-8278-bfb5551ad2f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.970326 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw5s4\" (UniqueName: \"kubernetes.io/projected/2dda26c0-0897-4f06-8278-bfb5551ad2f8-kube-api-access-tw5s4\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.970362 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:34 crc kubenswrapper[4575]: I1004 05:21:34.970371 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dda26c0-0897-4f06-8278-bfb5551ad2f8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.365104 4575 generic.go:334] "Generic (PLEG): container finished" podID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerID="58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f" exitCode=0 Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.365172 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7qzkh" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.365172 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qzkh" event={"ID":"2dda26c0-0897-4f06-8278-bfb5551ad2f8","Type":"ContainerDied","Data":"58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f"} Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.365260 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7qzkh" event={"ID":"2dda26c0-0897-4f06-8278-bfb5551ad2f8","Type":"ContainerDied","Data":"f18dbf83105cac8c45380deff22c904a0c6e34f2c756c7b21ec747c58ce030d1"} Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.365286 4575 scope.go:117] "RemoveContainer" containerID="58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.398195 4575 scope.go:117] "RemoveContainer" containerID="5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.407330 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7qzkh"] Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.433928 4575 scope.go:117] "RemoveContainer" containerID="2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.436698 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7qzkh"] Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.463600 4575 scope.go:117] "RemoveContainer" containerID="58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f" Oct 04 05:21:35 crc kubenswrapper[4575]: E1004 05:21:35.464032 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f\": container with ID starting with 58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f not found: ID does not exist" containerID="58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.464065 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f"} err="failed to get container status \"58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f\": rpc error: code = NotFound desc = could not find container \"58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f\": container with ID starting with 58102c4a0e8115d77df85371a64a822b31f9f3d2044cc60f9c0654366165167f not found: ID does not exist" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.464104 4575 scope.go:117] "RemoveContainer" containerID="5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3" Oct 04 05:21:35 crc kubenswrapper[4575]: E1004 05:21:35.464510 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3\": container with ID starting with 5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3 not found: ID does not exist" containerID="5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.464533 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3"} err="failed to get container status \"5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3\": rpc error: code = NotFound desc = could not find container \"5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3\": container with ID starting with 5cceaf41f2154f3856fcc6b583bbbb06d5e33ff65ff35938dbf4f2ff6b719df3 not found: ID does not exist" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.464551 4575 scope.go:117] "RemoveContainer" containerID="2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b" Oct 04 05:21:35 crc kubenswrapper[4575]: E1004 05:21:35.465737 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b\": container with ID starting with 2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b not found: ID does not exist" containerID="2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b" Oct 04 05:21:35 crc kubenswrapper[4575]: I1004 05:21:35.465785 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b"} err="failed to get container status \"2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b\": rpc error: code = NotFound desc = could not find container \"2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b\": container with ID starting with 2624523f6c1b66a668c4db7accb99b899026c394b1b5d8b78f923af9bd22766b not found: ID does not exist" Oct 04 05:21:37 crc kubenswrapper[4575]: I1004 05:21:37.322066 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" path="/var/lib/kubelet/pods/2dda26c0-0897-4f06-8278-bfb5551ad2f8/volumes" Oct 04 05:23:08 crc kubenswrapper[4575]: I1004 05:23:08.446108 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:23:08 crc kubenswrapper[4575]: I1004 05:23:08.448197 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.289069 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgm"] Oct 04 05:23:11 crc kubenswrapper[4575]: E1004 05:23:11.290026 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerName="extract-content" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.290042 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerName="extract-content" Oct 04 05:23:11 crc kubenswrapper[4575]: E1004 05:23:11.290062 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerName="extract-utilities" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.290069 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerName="extract-utilities" Oct 04 05:23:11 crc kubenswrapper[4575]: E1004 05:23:11.290089 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerName="registry-server" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.290096 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerName="registry-server" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.290268 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dda26c0-0897-4f06-8278-bfb5551ad2f8" containerName="registry-server" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.291730 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.321887 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgm"] Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.442565 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-catalog-content\") pod \"redhat-marketplace-nztgm\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.442722 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5kc2\" (UniqueName: \"kubernetes.io/projected/a35761fb-15d5-4821-ab14-f98a3a02d8f9-kube-api-access-m5kc2\") pod \"redhat-marketplace-nztgm\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.442751 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-utilities\") pod \"redhat-marketplace-nztgm\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.544972 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-catalog-content\") pod \"redhat-marketplace-nztgm\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.545084 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5kc2\" (UniqueName: \"kubernetes.io/projected/a35761fb-15d5-4821-ab14-f98a3a02d8f9-kube-api-access-m5kc2\") pod \"redhat-marketplace-nztgm\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.545119 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-utilities\") pod \"redhat-marketplace-nztgm\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.546270 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-catalog-content\") pod \"redhat-marketplace-nztgm\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.546574 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-utilities\") pod \"redhat-marketplace-nztgm\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.570188 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5kc2\" (UniqueName: \"kubernetes.io/projected/a35761fb-15d5-4821-ab14-f98a3a02d8f9-kube-api-access-m5kc2\") pod \"redhat-marketplace-nztgm\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:11 crc kubenswrapper[4575]: I1004 05:23:11.614732 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:12 crc kubenswrapper[4575]: I1004 05:23:12.150869 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgm"] Oct 04 05:23:12 crc kubenswrapper[4575]: I1004 05:23:12.248439 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgm" event={"ID":"a35761fb-15d5-4821-ab14-f98a3a02d8f9","Type":"ContainerStarted","Data":"2337dd9e44a95e4dca40e21fca9098d5fda041436ebe02315093fb3eccf6c2f9"} Oct 04 05:23:13 crc kubenswrapper[4575]: I1004 05:23:13.261554 4575 generic.go:334] "Generic (PLEG): container finished" podID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerID="af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471" exitCode=0 Oct 04 05:23:13 crc kubenswrapper[4575]: I1004 05:23:13.261632 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgm" event={"ID":"a35761fb-15d5-4821-ab14-f98a3a02d8f9","Type":"ContainerDied","Data":"af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471"} Oct 04 05:23:15 crc kubenswrapper[4575]: I1004 05:23:15.282177 4575 generic.go:334] "Generic (PLEG): container finished" podID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerID="053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7" exitCode=0 Oct 04 05:23:15 crc kubenswrapper[4575]: I1004 05:23:15.282337 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgm" event={"ID":"a35761fb-15d5-4821-ab14-f98a3a02d8f9","Type":"ContainerDied","Data":"053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7"} Oct 04 05:23:16 crc kubenswrapper[4575]: I1004 05:23:16.294025 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgm" event={"ID":"a35761fb-15d5-4821-ab14-f98a3a02d8f9","Type":"ContainerStarted","Data":"e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39"} Oct 04 05:23:16 crc kubenswrapper[4575]: I1004 05:23:16.322184 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nztgm" podStartSLOduration=2.891701952 podStartE2EDuration="5.322164808s" podCreationTimestamp="2025-10-04 05:23:11 +0000 UTC" firstStartedPulling="2025-10-04 05:23:13.263737522 +0000 UTC m=+2944.592296336" lastFinishedPulling="2025-10-04 05:23:15.694200378 +0000 UTC m=+2947.022759192" observedRunningTime="2025-10-04 05:23:16.315839285 +0000 UTC m=+2947.644398119" watchObservedRunningTime="2025-10-04 05:23:16.322164808 +0000 UTC m=+2947.650723622" Oct 04 05:23:21 crc kubenswrapper[4575]: I1004 05:23:21.615754 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:21 crc kubenswrapper[4575]: I1004 05:23:21.616400 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:21 crc kubenswrapper[4575]: I1004 05:23:21.670788 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:22 crc kubenswrapper[4575]: I1004 05:23:22.416659 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:22 crc kubenswrapper[4575]: I1004 05:23:22.460197 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgm"] Oct 04 05:23:24 crc kubenswrapper[4575]: I1004 05:23:24.389310 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nztgm" podUID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerName="registry-server" containerID="cri-o://e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39" gracePeriod=2 Oct 04 05:23:24 crc kubenswrapper[4575]: I1004 05:23:24.868932 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:24 crc kubenswrapper[4575]: I1004 05:23:24.996082 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5kc2\" (UniqueName: \"kubernetes.io/projected/a35761fb-15d5-4821-ab14-f98a3a02d8f9-kube-api-access-m5kc2\") pod \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " Oct 04 05:23:24 crc kubenswrapper[4575]: I1004 05:23:24.996484 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-catalog-content\") pod \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " Oct 04 05:23:24 crc kubenswrapper[4575]: I1004 05:23:24.996602 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-utilities\") pod \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\" (UID: \"a35761fb-15d5-4821-ab14-f98a3a02d8f9\") " Oct 04 05:23:24 crc kubenswrapper[4575]: I1004 05:23:24.997340 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-utilities" (OuterVolumeSpecName: "utilities") pod "a35761fb-15d5-4821-ab14-f98a3a02d8f9" (UID: "a35761fb-15d5-4821-ab14-f98a3a02d8f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.002481 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a35761fb-15d5-4821-ab14-f98a3a02d8f9-kube-api-access-m5kc2" (OuterVolumeSpecName: "kube-api-access-m5kc2") pod "a35761fb-15d5-4821-ab14-f98a3a02d8f9" (UID: "a35761fb-15d5-4821-ab14-f98a3a02d8f9"). InnerVolumeSpecName "kube-api-access-m5kc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.010663 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a35761fb-15d5-4821-ab14-f98a3a02d8f9" (UID: "a35761fb-15d5-4821-ab14-f98a3a02d8f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.098902 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.098944 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5kc2\" (UniqueName: \"kubernetes.io/projected/a35761fb-15d5-4821-ab14-f98a3a02d8f9-kube-api-access-m5kc2\") on node \"crc\" DevicePath \"\"" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.098957 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a35761fb-15d5-4821-ab14-f98a3a02d8f9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.400212 4575 generic.go:334] "Generic (PLEG): container finished" podID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerID="e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39" exitCode=0 Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.400267 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgm" event={"ID":"a35761fb-15d5-4821-ab14-f98a3a02d8f9","Type":"ContainerDied","Data":"e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39"} Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.400303 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgm" event={"ID":"a35761fb-15d5-4821-ab14-f98a3a02d8f9","Type":"ContainerDied","Data":"2337dd9e44a95e4dca40e21fca9098d5fda041436ebe02315093fb3eccf6c2f9"} Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.400329 4575 scope.go:117] "RemoveContainer" containerID="e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.400330 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nztgm" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.438491 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgm"] Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.447523 4575 scope.go:117] "RemoveContainer" containerID="053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.452629 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgm"] Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.465909 4575 scope.go:117] "RemoveContainer" containerID="af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.520578 4575 scope.go:117] "RemoveContainer" containerID="e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39" Oct 04 05:23:25 crc kubenswrapper[4575]: E1004 05:23:25.521050 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39\": container with ID starting with e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39 not found: ID does not exist" containerID="e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.521088 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39"} err="failed to get container status \"e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39\": rpc error: code = NotFound desc = could not find container \"e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39\": container with ID starting with e8b0cc937e663d5ae66b8e34855341f335af1ac3edd1bd45a485df7cccc9db39 not found: ID does not exist" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.521114 4575 scope.go:117] "RemoveContainer" containerID="053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7" Oct 04 05:23:25 crc kubenswrapper[4575]: E1004 05:23:25.521346 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7\": container with ID starting with 053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7 not found: ID does not exist" containerID="053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.521373 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7"} err="failed to get container status \"053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7\": rpc error: code = NotFound desc = could not find container \"053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7\": container with ID starting with 053919ed4a2bd90d1abaaaf0b1751f72fabc65fee448fc02a2667bd2380e19b7 not found: ID does not exist" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.521396 4575 scope.go:117] "RemoveContainer" containerID="af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471" Oct 04 05:23:25 crc kubenswrapper[4575]: E1004 05:23:25.521907 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471\": container with ID starting with af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471 not found: ID does not exist" containerID="af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471" Oct 04 05:23:25 crc kubenswrapper[4575]: I1004 05:23:25.521933 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471"} err="failed to get container status \"af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471\": rpc error: code = NotFound desc = could not find container \"af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471\": container with ID starting with af98176be2c3970920cda163a07e0f0088e347790c4260e5e97396fb6c70b471 not found: ID does not exist" Oct 04 05:23:27 crc kubenswrapper[4575]: I1004 05:23:27.322888 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" path="/var/lib/kubelet/pods/a35761fb-15d5-4821-ab14-f98a3a02d8f9/volumes" Oct 04 05:23:38 crc kubenswrapper[4575]: I1004 05:23:38.445909 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:23:38 crc kubenswrapper[4575]: I1004 05:23:38.446729 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:24:08 crc kubenswrapper[4575]: I1004 05:24:08.446166 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:24:08 crc kubenswrapper[4575]: I1004 05:24:08.446773 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:24:08 crc kubenswrapper[4575]: I1004 05:24:08.446817 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:24:08 crc kubenswrapper[4575]: I1004 05:24:08.447677 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:24:08 crc kubenswrapper[4575]: I1004 05:24:08.447741 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" gracePeriod=600 Oct 04 05:24:08 crc kubenswrapper[4575]: E1004 05:24:08.575268 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:24:08 crc kubenswrapper[4575]: I1004 05:24:08.793639 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" exitCode=0 Oct 04 05:24:08 crc kubenswrapper[4575]: I1004 05:24:08.793708 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26"} Oct 04 05:24:08 crc kubenswrapper[4575]: I1004 05:24:08.793769 4575 scope.go:117] "RemoveContainer" containerID="b77dbadefa0a5842cfdc96afce016aff88e8e24fb9b590a5ca3d5e7bcec7c7aa" Oct 04 05:24:08 crc kubenswrapper[4575]: I1004 05:24:08.794978 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:24:08 crc kubenswrapper[4575]: E1004 05:24:08.795522 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:24:22 crc kubenswrapper[4575]: I1004 05:24:22.310984 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:24:22 crc kubenswrapper[4575]: E1004 05:24:22.313834 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:24:37 crc kubenswrapper[4575]: I1004 05:24:37.310321 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:24:37 crc kubenswrapper[4575]: E1004 05:24:37.311315 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:24:51 crc kubenswrapper[4575]: I1004 05:24:51.216041 4575 generic.go:334] "Generic (PLEG): container finished" podID="fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" containerID="b4a62906508c4f5bed04c43401e4f2c71f6644eab6768ad65387bc0e0da7d5e2" exitCode=0 Oct 04 05:24:51 crc kubenswrapper[4575]: I1004 05:24:51.216147 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" event={"ID":"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17","Type":"ContainerDied","Data":"b4a62906508c4f5bed04c43401e4f2c71f6644eab6768ad65387bc0e0da7d5e2"} Oct 04 05:24:51 crc kubenswrapper[4575]: I1004 05:24:51.310702 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:24:51 crc kubenswrapper[4575]: E1004 05:24:51.311645 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.638252 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.657071 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-1\") pod \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.657138 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-1\") pod \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.657169 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-ssh-key\") pod \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.657287 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-inventory\") pod \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.657398 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-0\") pod \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.657491 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-extra-config-0\") pod \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.657570 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-combined-ca-bundle\") pod \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.657702 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-0\") pod \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.657752 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mftjs\" (UniqueName: \"kubernetes.io/projected/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-kube-api-access-mftjs\") pod \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\" (UID: \"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17\") " Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.668982 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-kube-api-access-mftjs" (OuterVolumeSpecName: "kube-api-access-mftjs") pod "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" (UID: "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17"). InnerVolumeSpecName "kube-api-access-mftjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.698287 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" (UID: "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.709626 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" (UID: "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.724385 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" (UID: "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.736568 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" (UID: "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.740837 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" (UID: "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.742002 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" (UID: "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.744385 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" (UID: "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.755171 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-inventory" (OuterVolumeSpecName: "inventory") pod "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" (UID: "fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.759446 4575 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.759669 4575 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.759750 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mftjs\" (UniqueName: \"kubernetes.io/projected/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-kube-api-access-mftjs\") on node \"crc\" DevicePath \"\"" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.759829 4575 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.759910 4575 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.759997 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.760084 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.760162 4575 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:24:52 crc kubenswrapper[4575]: I1004 05:24:52.760280 4575 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.237334 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" event={"ID":"fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17","Type":"ContainerDied","Data":"84af6207453a50e6bb95146d5972c4862974b300fc75f6524c1187e6525f060e"} Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.237385 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ft442" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.237397 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84af6207453a50e6bb95146d5972c4862974b300fc75f6524c1187e6525f060e" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.351869 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc"] Oct 04 05:24:53 crc kubenswrapper[4575]: E1004 05:24:53.352382 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerName="extract-utilities" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.352408 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerName="extract-utilities" Oct 04 05:24:53 crc kubenswrapper[4575]: E1004 05:24:53.352445 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerName="extract-content" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.352454 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerName="extract-content" Oct 04 05:24:53 crc kubenswrapper[4575]: E1004 05:24:53.352481 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerName="registry-server" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.352491 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerName="registry-server" Oct 04 05:24:53 crc kubenswrapper[4575]: E1004 05:24:53.352512 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.352521 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.352757 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="a35761fb-15d5-4821-ab14-f98a3a02d8f9" containerName="registry-server" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.352812 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.353565 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.355945 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.356504 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.356817 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.356998 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.357160 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fxszm" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.372204 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc"] Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.374130 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.374298 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.374339 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.374364 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.374421 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lltf6\" (UniqueName: \"kubernetes.io/projected/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-kube-api-access-lltf6\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.374446 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.374540 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.476091 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.476283 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.476318 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.476375 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lltf6\" (UniqueName: \"kubernetes.io/projected/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-kube-api-access-lltf6\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.476415 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.476495 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.476552 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.480974 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.481801 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.482289 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.482860 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.484466 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.484544 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.496041 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lltf6\" (UniqueName: \"kubernetes.io/projected/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-kube-api-access-lltf6\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:53 crc kubenswrapper[4575]: I1004 05:24:53.719097 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:24:54 crc kubenswrapper[4575]: I1004 05:24:54.287961 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc"] Oct 04 05:24:55 crc kubenswrapper[4575]: I1004 05:24:55.257545 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" event={"ID":"f17c12b6-fe62-4659-ba3f-e61d9a1cac38","Type":"ContainerStarted","Data":"715a507240e30c0bc49f6242d30aaabd07d310e75d9c22c8bf08955e30385ff0"} Oct 04 05:24:55 crc kubenswrapper[4575]: I1004 05:24:55.257947 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" event={"ID":"f17c12b6-fe62-4659-ba3f-e61d9a1cac38","Type":"ContainerStarted","Data":"5a65d1abbe4d16329a82ae95c0a0f7f98b9afab8885e95e08e7032b0e191a690"} Oct 04 05:25:02 crc kubenswrapper[4575]: I1004 05:25:02.310779 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:25:02 crc kubenswrapper[4575]: E1004 05:25:02.312023 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:25:16 crc kubenswrapper[4575]: I1004 05:25:16.310741 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:25:16 crc kubenswrapper[4575]: E1004 05:25:16.312739 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:25:30 crc kubenswrapper[4575]: I1004 05:25:30.311088 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:25:30 crc kubenswrapper[4575]: E1004 05:25:30.312328 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:25:43 crc kubenswrapper[4575]: I1004 05:25:43.311243 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:25:43 crc kubenswrapper[4575]: E1004 05:25:43.311947 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:25:58 crc kubenswrapper[4575]: I1004 05:25:58.311689 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:25:58 crc kubenswrapper[4575]: E1004 05:25:58.312405 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:26:12 crc kubenswrapper[4575]: I1004 05:26:12.310876 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:26:12 crc kubenswrapper[4575]: E1004 05:26:12.311721 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:26:25 crc kubenswrapper[4575]: I1004 05:26:25.310217 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:26:25 crc kubenswrapper[4575]: E1004 05:26:25.311228 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.232428 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" podStartSLOduration=98.813444566 podStartE2EDuration="1m39.232404625s" podCreationTimestamp="2025-10-04 05:24:53 +0000 UTC" firstStartedPulling="2025-10-04 05:24:54.290132538 +0000 UTC m=+3045.618691352" lastFinishedPulling="2025-10-04 05:24:54.709092607 +0000 UTC m=+3046.037651411" observedRunningTime="2025-10-04 05:24:55.283634895 +0000 UTC m=+3046.612193719" watchObservedRunningTime="2025-10-04 05:26:32.232404625 +0000 UTC m=+3143.560963439" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.247570 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gd5s6"] Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.251681 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.264259 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gd5s6"] Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.360607 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20397c5b-8ea1-4ffe-851f-d88f1b228967-catalog-content\") pod \"community-operators-gd5s6\" (UID: \"20397c5b-8ea1-4ffe-851f-d88f1b228967\") " pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.360742 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20397c5b-8ea1-4ffe-851f-d88f1b228967-utilities\") pod \"community-operators-gd5s6\" (UID: \"20397c5b-8ea1-4ffe-851f-d88f1b228967\") " pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.360800 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvnxb\" (UniqueName: \"kubernetes.io/projected/20397c5b-8ea1-4ffe-851f-d88f1b228967-kube-api-access-pvnxb\") pod \"community-operators-gd5s6\" (UID: \"20397c5b-8ea1-4ffe-851f-d88f1b228967\") " pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.462468 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20397c5b-8ea1-4ffe-851f-d88f1b228967-catalog-content\") pod \"community-operators-gd5s6\" (UID: \"20397c5b-8ea1-4ffe-851f-d88f1b228967\") " pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.462599 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20397c5b-8ea1-4ffe-851f-d88f1b228967-utilities\") pod \"community-operators-gd5s6\" (UID: \"20397c5b-8ea1-4ffe-851f-d88f1b228967\") " pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.462673 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvnxb\" (UniqueName: \"kubernetes.io/projected/20397c5b-8ea1-4ffe-851f-d88f1b228967-kube-api-access-pvnxb\") pod \"community-operators-gd5s6\" (UID: \"20397c5b-8ea1-4ffe-851f-d88f1b228967\") " pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.463631 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20397c5b-8ea1-4ffe-851f-d88f1b228967-catalog-content\") pod \"community-operators-gd5s6\" (UID: \"20397c5b-8ea1-4ffe-851f-d88f1b228967\") " pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.463923 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20397c5b-8ea1-4ffe-851f-d88f1b228967-utilities\") pod \"community-operators-gd5s6\" (UID: \"20397c5b-8ea1-4ffe-851f-d88f1b228967\") " pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.489808 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvnxb\" (UniqueName: \"kubernetes.io/projected/20397c5b-8ea1-4ffe-851f-d88f1b228967-kube-api-access-pvnxb\") pod \"community-operators-gd5s6\" (UID: \"20397c5b-8ea1-4ffe-851f-d88f1b228967\") " pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:32 crc kubenswrapper[4575]: I1004 05:26:32.577190 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:33 crc kubenswrapper[4575]: I1004 05:26:33.218090 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gd5s6"] Oct 04 05:26:34 crc kubenswrapper[4575]: I1004 05:26:34.145900 4575 generic.go:334] "Generic (PLEG): container finished" podID="20397c5b-8ea1-4ffe-851f-d88f1b228967" containerID="90fa03179f58b58276d47edb86f60db259bf3c5f2aba2e4f93694c1e9c4c0129" exitCode=0 Oct 04 05:26:34 crc kubenswrapper[4575]: I1004 05:26:34.145999 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5s6" event={"ID":"20397c5b-8ea1-4ffe-851f-d88f1b228967","Type":"ContainerDied","Data":"90fa03179f58b58276d47edb86f60db259bf3c5f2aba2e4f93694c1e9c4c0129"} Oct 04 05:26:34 crc kubenswrapper[4575]: I1004 05:26:34.147935 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5s6" event={"ID":"20397c5b-8ea1-4ffe-851f-d88f1b228967","Type":"ContainerStarted","Data":"8abc37f8b5b3fc82e162b438a3ff96b1d9a0dc6d5d398b3ed942f9a1eec38f9a"} Oct 04 05:26:34 crc kubenswrapper[4575]: I1004 05:26:34.149699 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:26:36 crc kubenswrapper[4575]: I1004 05:26:36.311389 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:26:36 crc kubenswrapper[4575]: E1004 05:26:36.312240 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:26:40 crc kubenswrapper[4575]: I1004 05:26:40.227036 4575 generic.go:334] "Generic (PLEG): container finished" podID="20397c5b-8ea1-4ffe-851f-d88f1b228967" containerID="2411ccda877b46e23c9ed84db28b4f5f071f2740685f63b97c5afb39d47b7df7" exitCode=0 Oct 04 05:26:40 crc kubenswrapper[4575]: I1004 05:26:40.227259 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5s6" event={"ID":"20397c5b-8ea1-4ffe-851f-d88f1b228967","Type":"ContainerDied","Data":"2411ccda877b46e23c9ed84db28b4f5f071f2740685f63b97c5afb39d47b7df7"} Oct 04 05:26:41 crc kubenswrapper[4575]: I1004 05:26:41.245291 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5s6" event={"ID":"20397c5b-8ea1-4ffe-851f-d88f1b228967","Type":"ContainerStarted","Data":"49c8281f4905495cf277b0a1d218c09580c7bf6a466a6ec53049c7c3edfcd7d4"} Oct 04 05:26:41 crc kubenswrapper[4575]: I1004 05:26:41.271423 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gd5s6" podStartSLOduration=2.760178664 podStartE2EDuration="9.271400972s" podCreationTimestamp="2025-10-04 05:26:32 +0000 UTC" firstStartedPulling="2025-10-04 05:26:34.149415 +0000 UTC m=+3145.477973814" lastFinishedPulling="2025-10-04 05:26:40.660637308 +0000 UTC m=+3151.989196122" observedRunningTime="2025-10-04 05:26:41.264191624 +0000 UTC m=+3152.592750448" watchObservedRunningTime="2025-10-04 05:26:41.271400972 +0000 UTC m=+3152.599959786" Oct 04 05:26:42 crc kubenswrapper[4575]: I1004 05:26:42.577987 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:42 crc kubenswrapper[4575]: I1004 05:26:42.579418 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:43 crc kubenswrapper[4575]: I1004 05:26:43.633319 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gd5s6" podUID="20397c5b-8ea1-4ffe-851f-d88f1b228967" containerName="registry-server" probeResult="failure" output=< Oct 04 05:26:43 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:26:43 crc kubenswrapper[4575]: > Oct 04 05:26:51 crc kubenswrapper[4575]: I1004 05:26:51.310164 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:26:51 crc kubenswrapper[4575]: E1004 05:26:51.310954 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:26:52 crc kubenswrapper[4575]: I1004 05:26:52.628342 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:52 crc kubenswrapper[4575]: I1004 05:26:52.685118 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gd5s6" Oct 04 05:26:52 crc kubenswrapper[4575]: I1004 05:26:52.791633 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gd5s6"] Oct 04 05:26:52 crc kubenswrapper[4575]: I1004 05:26:52.870980 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ktd5v"] Oct 04 05:26:52 crc kubenswrapper[4575]: I1004 05:26:52.871374 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ktd5v" podUID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerName="registry-server" containerID="cri-o://d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac" gracePeriod=2 Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.348987 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ktd5v" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.357272 4575 generic.go:334] "Generic (PLEG): container finished" podID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerID="d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac" exitCode=0 Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.358291 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ktd5v" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.358460 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktd5v" event={"ID":"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354","Type":"ContainerDied","Data":"d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac"} Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.358489 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktd5v" event={"ID":"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354","Type":"ContainerDied","Data":"800ee2598d3d07d51d233d407952e1cb49ca30be90e05cc5bcf2f25d29b6f75a"} Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.358508 4575 scope.go:117] "RemoveContainer" containerID="d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.385253 4575 scope.go:117] "RemoveContainer" containerID="314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.387238 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-catalog-content\") pod \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.387323 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzr77\" (UniqueName: \"kubernetes.io/projected/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-kube-api-access-nzr77\") pod \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.387475 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-utilities\") pod \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\" (UID: \"c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354\") " Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.388297 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-utilities" (OuterVolumeSpecName: "utilities") pod "c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" (UID: "c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.389317 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.413023 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-kube-api-access-nzr77" (OuterVolumeSpecName: "kube-api-access-nzr77") pod "c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" (UID: "c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354"). InnerVolumeSpecName "kube-api-access-nzr77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.443892 4575 scope.go:117] "RemoveContainer" containerID="b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.492035 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzr77\" (UniqueName: \"kubernetes.io/projected/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-kube-api-access-nzr77\") on node \"crc\" DevicePath \"\"" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.495457 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" (UID: "c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.510866 4575 scope.go:117] "RemoveContainer" containerID="d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac" Oct 04 05:26:53 crc kubenswrapper[4575]: E1004 05:26:53.512897 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac\": container with ID starting with d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac not found: ID does not exist" containerID="d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.512934 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac"} err="failed to get container status \"d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac\": rpc error: code = NotFound desc = could not find container \"d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac\": container with ID starting with d502ecc69006da0552f363e7f50727b50d25db79289ed11550791223f70c16ac not found: ID does not exist" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.512955 4575 scope.go:117] "RemoveContainer" containerID="314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda" Oct 04 05:26:53 crc kubenswrapper[4575]: E1004 05:26:53.513419 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda\": container with ID starting with 314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda not found: ID does not exist" containerID="314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.513467 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda"} err="failed to get container status \"314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda\": rpc error: code = NotFound desc = could not find container \"314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda\": container with ID starting with 314e1f93e6542352817ee405577f6ade6a0213aded41413d831843e2e32dbbda not found: ID does not exist" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.513493 4575 scope.go:117] "RemoveContainer" containerID="b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0" Oct 04 05:26:53 crc kubenswrapper[4575]: E1004 05:26:53.513894 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0\": container with ID starting with b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0 not found: ID does not exist" containerID="b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.513922 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0"} err="failed to get container status \"b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0\": rpc error: code = NotFound desc = could not find container \"b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0\": container with ID starting with b1029da24f2046749137a80be01c3f84c64ca119dea9c167a5e0b0ae0949e7d0 not found: ID does not exist" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.594181 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.690245 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ktd5v"] Oct 04 05:26:53 crc kubenswrapper[4575]: I1004 05:26:53.699512 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ktd5v"] Oct 04 05:26:55 crc kubenswrapper[4575]: I1004 05:26:55.320806 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" path="/var/lib/kubelet/pods/c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354/volumes" Oct 04 05:27:05 crc kubenswrapper[4575]: I1004 05:27:05.310237 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:27:05 crc kubenswrapper[4575]: E1004 05:27:05.311002 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:27:20 crc kubenswrapper[4575]: I1004 05:27:20.310426 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:27:20 crc kubenswrapper[4575]: E1004 05:27:20.311956 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:27:34 crc kubenswrapper[4575]: I1004 05:27:34.311277 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:27:34 crc kubenswrapper[4575]: E1004 05:27:34.312306 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:27:47 crc kubenswrapper[4575]: I1004 05:27:47.309806 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:27:47 crc kubenswrapper[4575]: E1004 05:27:47.310561 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:27:58 crc kubenswrapper[4575]: I1004 05:27:58.310447 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:27:58 crc kubenswrapper[4575]: E1004 05:27:58.311230 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:28:12 crc kubenswrapper[4575]: I1004 05:28:12.309800 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:28:12 crc kubenswrapper[4575]: E1004 05:28:12.310611 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:28:23 crc kubenswrapper[4575]: I1004 05:28:23.311058 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:28:23 crc kubenswrapper[4575]: E1004 05:28:23.312011 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:28:26 crc kubenswrapper[4575]: I1004 05:28:26.162422 4575 generic.go:334] "Generic (PLEG): container finished" podID="f17c12b6-fe62-4659-ba3f-e61d9a1cac38" containerID="715a507240e30c0bc49f6242d30aaabd07d310e75d9c22c8bf08955e30385ff0" exitCode=0 Oct 04 05:28:26 crc kubenswrapper[4575]: I1004 05:28:26.162513 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" event={"ID":"f17c12b6-fe62-4659-ba3f-e61d9a1cac38","Type":"ContainerDied","Data":"715a507240e30c0bc49f6242d30aaabd07d310e75d9c22c8bf08955e30385ff0"} Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.625838 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.686729 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-telemetry-combined-ca-bundle\") pod \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.686825 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ssh-key\") pod \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.686891 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-0\") pod \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.687034 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lltf6\" (UniqueName: \"kubernetes.io/projected/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-kube-api-access-lltf6\") pod \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.687906 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-inventory\") pod \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.688028 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-2\") pod \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.688067 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-1\") pod \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\" (UID: \"f17c12b6-fe62-4659-ba3f-e61d9a1cac38\") " Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.693380 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f17c12b6-fe62-4659-ba3f-e61d9a1cac38" (UID: "f17c12b6-fe62-4659-ba3f-e61d9a1cac38"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.694054 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-kube-api-access-lltf6" (OuterVolumeSpecName: "kube-api-access-lltf6") pod "f17c12b6-fe62-4659-ba3f-e61d9a1cac38" (UID: "f17c12b6-fe62-4659-ba3f-e61d9a1cac38"). InnerVolumeSpecName "kube-api-access-lltf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.717721 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-inventory" (OuterVolumeSpecName: "inventory") pod "f17c12b6-fe62-4659-ba3f-e61d9a1cac38" (UID: "f17c12b6-fe62-4659-ba3f-e61d9a1cac38"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.719708 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f17c12b6-fe62-4659-ba3f-e61d9a1cac38" (UID: "f17c12b6-fe62-4659-ba3f-e61d9a1cac38"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.724212 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "f17c12b6-fe62-4659-ba3f-e61d9a1cac38" (UID: "f17c12b6-fe62-4659-ba3f-e61d9a1cac38"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.724355 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "f17c12b6-fe62-4659-ba3f-e61d9a1cac38" (UID: "f17c12b6-fe62-4659-ba3f-e61d9a1cac38"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.732617 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "f17c12b6-fe62-4659-ba3f-e61d9a1cac38" (UID: "f17c12b6-fe62-4659-ba3f-e61d9a1cac38"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.790525 4575 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.790577 4575 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.790606 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.790619 4575 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.790632 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lltf6\" (UniqueName: \"kubernetes.io/projected/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-kube-api-access-lltf6\") on node \"crc\" DevicePath \"\"" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.790649 4575 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:28:27 crc kubenswrapper[4575]: I1004 05:28:27.790663 4575 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f17c12b6-fe62-4659-ba3f-e61d9a1cac38-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 04 05:28:28 crc kubenswrapper[4575]: I1004 05:28:28.183080 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" Oct 04 05:28:28 crc kubenswrapper[4575]: I1004 05:28:28.182973 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc" event={"ID":"f17c12b6-fe62-4659-ba3f-e61d9a1cac38","Type":"ContainerDied","Data":"5a65d1abbe4d16329a82ae95c0a0f7f98b9afab8885e95e08e7032b0e191a690"} Oct 04 05:28:28 crc kubenswrapper[4575]: I1004 05:28:28.183730 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a65d1abbe4d16329a82ae95c0a0f7f98b9afab8885e95e08e7032b0e191a690" Oct 04 05:28:36 crc kubenswrapper[4575]: I1004 05:28:36.310498 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:28:36 crc kubenswrapper[4575]: E1004 05:28:36.311357 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:28:49 crc kubenswrapper[4575]: I1004 05:28:49.321143 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:28:49 crc kubenswrapper[4575]: E1004 05:28:49.322356 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:29:00 crc kubenswrapper[4575]: I1004 05:29:00.310845 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:29:00 crc kubenswrapper[4575]: E1004 05:29:00.311667 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:29:15 crc kubenswrapper[4575]: I1004 05:29:15.310465 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:29:15 crc kubenswrapper[4575]: I1004 05:29:15.598880 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"c2adda20348c666e32616a3a23aff161d88262e97b92e05c789aa15c41b2f3d1"} Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.001431 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 05:29:19 crc kubenswrapper[4575]: E1004 05:29:19.002581 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerName="registry-server" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.002620 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerName="registry-server" Oct 04 05:29:19 crc kubenswrapper[4575]: E1004 05:29:19.002656 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17c12b6-fe62-4659-ba3f-e61d9a1cac38" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.002665 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17c12b6-fe62-4659-ba3f-e61d9a1cac38" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 05:29:19 crc kubenswrapper[4575]: E1004 05:29:19.002689 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerName="extract-utilities" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.002696 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerName="extract-utilities" Oct 04 05:29:19 crc kubenswrapper[4575]: E1004 05:29:19.002719 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerName="extract-content" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.002726 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerName="extract-content" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.002897 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6b72fdb-cfb9-4bb7-bb8e-2ffbfbbb8354" containerName="registry-server" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.002924 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17c12b6-fe62-4659-ba3f-e61d9a1cac38" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.004279 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.006719 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.007042 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.007876 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.007897 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9mw2d" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.013646 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.031274 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.033896 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-config-data\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.034104 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.136369 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.136434 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.136464 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-config-data\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.136487 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.136514 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.136554 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.136571 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.136656 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.136683 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk5w8\" (UniqueName: \"kubernetes.io/projected/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-kube-api-access-fk5w8\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.138197 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.138365 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-config-data\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.148333 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.238993 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.239062 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.239111 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.239521 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.239722 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.240506 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.240578 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk5w8\" (UniqueName: \"kubernetes.io/projected/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-kube-api-access-fk5w8\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.240593 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.240918 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.243978 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.244031 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.258453 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk5w8\" (UniqueName: \"kubernetes.io/projected/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-kube-api-access-fk5w8\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.274918 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.325299 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 05:29:19 crc kubenswrapper[4575]: I1004 05:29:19.776835 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 05:29:19 crc kubenswrapper[4575]: W1004 05:29:19.777220 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod339c2df5_cdf1_4c78_958e_9a4e65e1e2a7.slice/crio-17fbea7b37bbb9a17ed534d1325cb73c384fc832258ea9b76702cded3d313d06 WatchSource:0}: Error finding container 17fbea7b37bbb9a17ed534d1325cb73c384fc832258ea9b76702cded3d313d06: Status 404 returned error can't find the container with id 17fbea7b37bbb9a17ed534d1325cb73c384fc832258ea9b76702cded3d313d06 Oct 04 05:29:20 crc kubenswrapper[4575]: I1004 05:29:20.646503 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7","Type":"ContainerStarted","Data":"17fbea7b37bbb9a17ed534d1325cb73c384fc832258ea9b76702cded3d313d06"} Oct 04 05:29:56 crc kubenswrapper[4575]: E1004 05:29:56.189779 4575 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 04 05:29:56 crc kubenswrapper[4575]: E1004 05:29:56.191688 4575 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fk5w8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(339c2df5-cdf1-4c78-958e-9a4e65e1e2a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 05:29:56 crc kubenswrapper[4575]: E1004 05:29:56.195112 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" Oct 04 05:29:57 crc kubenswrapper[4575]: E1004 05:29:57.063224 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.174218 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc"] Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.176137 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.186109 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.186189 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.197295 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc"] Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.333292 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48fa75a3-f497-4a75-8153-54cf9ad31576-secret-volume\") pod \"collect-profiles-29325930-6dnvc\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.333936 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48fa75a3-f497-4a75-8153-54cf9ad31576-config-volume\") pod \"collect-profiles-29325930-6dnvc\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.334175 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfnt4\" (UniqueName: \"kubernetes.io/projected/48fa75a3-f497-4a75-8153-54cf9ad31576-kube-api-access-dfnt4\") pod \"collect-profiles-29325930-6dnvc\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.436137 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfnt4\" (UniqueName: \"kubernetes.io/projected/48fa75a3-f497-4a75-8153-54cf9ad31576-kube-api-access-dfnt4\") pod \"collect-profiles-29325930-6dnvc\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.436305 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48fa75a3-f497-4a75-8153-54cf9ad31576-secret-volume\") pod \"collect-profiles-29325930-6dnvc\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.436356 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48fa75a3-f497-4a75-8153-54cf9ad31576-config-volume\") pod \"collect-profiles-29325930-6dnvc\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.438378 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48fa75a3-f497-4a75-8153-54cf9ad31576-config-volume\") pod \"collect-profiles-29325930-6dnvc\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.446270 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48fa75a3-f497-4a75-8153-54cf9ad31576-secret-volume\") pod \"collect-profiles-29325930-6dnvc\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.457212 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfnt4\" (UniqueName: \"kubernetes.io/projected/48fa75a3-f497-4a75-8153-54cf9ad31576-kube-api-access-dfnt4\") pod \"collect-profiles-29325930-6dnvc\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:00 crc kubenswrapper[4575]: I1004 05:30:00.507776 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:01 crc kubenswrapper[4575]: I1004 05:30:01.028558 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc"] Oct 04 05:30:01 crc kubenswrapper[4575]: W1004 05:30:01.040792 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48fa75a3_f497_4a75_8153_54cf9ad31576.slice/crio-a03c820356598c7a4050ab7d13b3de539b0b22ad2b194f66c92a99d4483617e4 WatchSource:0}: Error finding container a03c820356598c7a4050ab7d13b3de539b0b22ad2b194f66c92a99d4483617e4: Status 404 returned error can't find the container with id a03c820356598c7a4050ab7d13b3de539b0b22ad2b194f66c92a99d4483617e4 Oct 04 05:30:01 crc kubenswrapper[4575]: I1004 05:30:01.110736 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" event={"ID":"48fa75a3-f497-4a75-8153-54cf9ad31576","Type":"ContainerStarted","Data":"a03c820356598c7a4050ab7d13b3de539b0b22ad2b194f66c92a99d4483617e4"} Oct 04 05:30:02 crc kubenswrapper[4575]: I1004 05:30:02.135093 4575 generic.go:334] "Generic (PLEG): container finished" podID="48fa75a3-f497-4a75-8153-54cf9ad31576" containerID="6e68f99175cee7251450df52e6fbd4789144875b7f8aa97ef441d17511abd54b" exitCode=0 Oct 04 05:30:02 crc kubenswrapper[4575]: I1004 05:30:02.135195 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" event={"ID":"48fa75a3-f497-4a75-8153-54cf9ad31576","Type":"ContainerDied","Data":"6e68f99175cee7251450df52e6fbd4789144875b7f8aa97ef441d17511abd54b"} Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.461980 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.608139 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfnt4\" (UniqueName: \"kubernetes.io/projected/48fa75a3-f497-4a75-8153-54cf9ad31576-kube-api-access-dfnt4\") pod \"48fa75a3-f497-4a75-8153-54cf9ad31576\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.608530 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48fa75a3-f497-4a75-8153-54cf9ad31576-config-volume\") pod \"48fa75a3-f497-4a75-8153-54cf9ad31576\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.608855 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48fa75a3-f497-4a75-8153-54cf9ad31576-secret-volume\") pod \"48fa75a3-f497-4a75-8153-54cf9ad31576\" (UID: \"48fa75a3-f497-4a75-8153-54cf9ad31576\") " Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.609178 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fa75a3-f497-4a75-8153-54cf9ad31576-config-volume" (OuterVolumeSpecName: "config-volume") pod "48fa75a3-f497-4a75-8153-54cf9ad31576" (UID: "48fa75a3-f497-4a75-8153-54cf9ad31576"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.609733 4575 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48fa75a3-f497-4a75-8153-54cf9ad31576-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.613867 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48fa75a3-f497-4a75-8153-54cf9ad31576-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "48fa75a3-f497-4a75-8153-54cf9ad31576" (UID: "48fa75a3-f497-4a75-8153-54cf9ad31576"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.624041 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48fa75a3-f497-4a75-8153-54cf9ad31576-kube-api-access-dfnt4" (OuterVolumeSpecName: "kube-api-access-dfnt4") pod "48fa75a3-f497-4a75-8153-54cf9ad31576" (UID: "48fa75a3-f497-4a75-8153-54cf9ad31576"). InnerVolumeSpecName "kube-api-access-dfnt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.711705 4575 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48fa75a3-f497-4a75-8153-54cf9ad31576-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:03 crc kubenswrapper[4575]: I1004 05:30:03.711742 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfnt4\" (UniqueName: \"kubernetes.io/projected/48fa75a3-f497-4a75-8153-54cf9ad31576-kube-api-access-dfnt4\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:04 crc kubenswrapper[4575]: I1004 05:30:04.152742 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" event={"ID":"48fa75a3-f497-4a75-8153-54cf9ad31576","Type":"ContainerDied","Data":"a03c820356598c7a4050ab7d13b3de539b0b22ad2b194f66c92a99d4483617e4"} Oct 04 05:30:04 crc kubenswrapper[4575]: I1004 05:30:04.153160 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a03c820356598c7a4050ab7d13b3de539b0b22ad2b194f66c92a99d4483617e4" Oct 04 05:30:04 crc kubenswrapper[4575]: I1004 05:30:04.152796 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-6dnvc" Oct 04 05:30:04 crc kubenswrapper[4575]: I1004 05:30:04.542806 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9"] Oct 04 05:30:04 crc kubenswrapper[4575]: I1004 05:30:04.553797 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-sl5k9"] Oct 04 05:30:05 crc kubenswrapper[4575]: I1004 05:30:05.325415 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f915159b-2f89-4709-a2f7-447ab38f7eaa" path="/var/lib/kubelet/pods/f915159b-2f89-4709-a2f7-447ab38f7eaa/volumes" Oct 04 05:30:13 crc kubenswrapper[4575]: I1004 05:30:13.022036 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 05:30:14 crc kubenswrapper[4575]: I1004 05:30:14.239622 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7","Type":"ContainerStarted","Data":"e52e54d3a67b8e66ee4c34579462f3980bda3212e22973adebbdafe81b7e60c8"} Oct 04 05:30:14 crc kubenswrapper[4575]: I1004 05:30:14.258151 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.019217027 podStartE2EDuration="57.2581289s" podCreationTimestamp="2025-10-04 05:29:17 +0000 UTC" firstStartedPulling="2025-10-04 05:29:19.780195082 +0000 UTC m=+3311.108753896" lastFinishedPulling="2025-10-04 05:30:13.019106955 +0000 UTC m=+3364.347665769" observedRunningTime="2025-10-04 05:30:14.257745719 +0000 UTC m=+3365.586304543" watchObservedRunningTime="2025-10-04 05:30:14.2581289 +0000 UTC m=+3365.586687724" Oct 04 05:30:31 crc kubenswrapper[4575]: I1004 05:30:31.169075 4575 scope.go:117] "RemoveContainer" containerID="9c8c0044b9785c42c82ed24dfcf93de0c1cc8268aad9a97747cd2243f7d33d34" Oct 04 05:31:38 crc kubenswrapper[4575]: I1004 05:31:38.445829 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:31:38 crc kubenswrapper[4575]: I1004 05:31:38.446790 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.452177 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bcrcn"] Oct 04 05:32:04 crc kubenswrapper[4575]: E1004 05:32:04.453633 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fa75a3-f497-4a75-8153-54cf9ad31576" containerName="collect-profiles" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.453655 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fa75a3-f497-4a75-8153-54cf9ad31576" containerName="collect-profiles" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.453908 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fa75a3-f497-4a75-8153-54cf9ad31576" containerName="collect-profiles" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.455627 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.479668 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bcrcn"] Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.501036 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-utilities\") pod \"certified-operators-bcrcn\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.501552 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhfgn\" (UniqueName: \"kubernetes.io/projected/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-kube-api-access-fhfgn\") pod \"certified-operators-bcrcn\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.501714 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-catalog-content\") pod \"certified-operators-bcrcn\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.603371 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhfgn\" (UniqueName: \"kubernetes.io/projected/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-kube-api-access-fhfgn\") pod \"certified-operators-bcrcn\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.603493 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-catalog-content\") pod \"certified-operators-bcrcn\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.603617 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-utilities\") pod \"certified-operators-bcrcn\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.604354 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-utilities\") pod \"certified-operators-bcrcn\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.604479 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-catalog-content\") pod \"certified-operators-bcrcn\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.629462 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhfgn\" (UniqueName: \"kubernetes.io/projected/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-kube-api-access-fhfgn\") pod \"certified-operators-bcrcn\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:04 crc kubenswrapper[4575]: I1004 05:32:04.784883 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:06 crc kubenswrapper[4575]: I1004 05:32:05.796412 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bcrcn"] Oct 04 05:32:06 crc kubenswrapper[4575]: I1004 05:32:06.372404 4575 generic.go:334] "Generic (PLEG): container finished" podID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerID="4b7fad6d46d7bccf33e5cb34d2e1be81f5d468579ce1fc9fb3564045974b46cf" exitCode=0 Oct 04 05:32:06 crc kubenswrapper[4575]: I1004 05:32:06.372930 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcrcn" event={"ID":"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb","Type":"ContainerDied","Data":"4b7fad6d46d7bccf33e5cb34d2e1be81f5d468579ce1fc9fb3564045974b46cf"} Oct 04 05:32:06 crc kubenswrapper[4575]: I1004 05:32:06.372977 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcrcn" event={"ID":"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb","Type":"ContainerStarted","Data":"9ddac1f0eb40651005f089648eadd7e45a5b536bf9682579b2841ac5d0204759"} Oct 04 05:32:06 crc kubenswrapper[4575]: I1004 05:32:06.377686 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:32:08 crc kubenswrapper[4575]: I1004 05:32:08.394316 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcrcn" event={"ID":"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb","Type":"ContainerStarted","Data":"95d28f711e1eef09df34a191ae492eccd899cebda29a737674857db7beeaa9cc"} Oct 04 05:32:08 crc kubenswrapper[4575]: I1004 05:32:08.446247 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:32:08 crc kubenswrapper[4575]: I1004 05:32:08.446312 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:32:14 crc kubenswrapper[4575]: I1004 05:32:14.455271 4575 generic.go:334] "Generic (PLEG): container finished" podID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerID="95d28f711e1eef09df34a191ae492eccd899cebda29a737674857db7beeaa9cc" exitCode=0 Oct 04 05:32:14 crc kubenswrapper[4575]: I1004 05:32:14.455467 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcrcn" event={"ID":"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb","Type":"ContainerDied","Data":"95d28f711e1eef09df34a191ae492eccd899cebda29a737674857db7beeaa9cc"} Oct 04 05:32:15 crc kubenswrapper[4575]: I1004 05:32:15.468445 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcrcn" event={"ID":"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb","Type":"ContainerStarted","Data":"9231e4e6ee5c1faab5b44f9131ceedb4c3e6cc5dfc632e43fc03f1d383b7a4f6"} Oct 04 05:32:15 crc kubenswrapper[4575]: I1004 05:32:15.485711 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bcrcn" podStartSLOduration=2.949563983 podStartE2EDuration="11.485696791s" podCreationTimestamp="2025-10-04 05:32:04 +0000 UTC" firstStartedPulling="2025-10-04 05:32:06.377421309 +0000 UTC m=+3477.705980123" lastFinishedPulling="2025-10-04 05:32:14.913554117 +0000 UTC m=+3486.242112931" observedRunningTime="2025-10-04 05:32:15.485217477 +0000 UTC m=+3486.813776291" watchObservedRunningTime="2025-10-04 05:32:15.485696791 +0000 UTC m=+3486.814255605" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.744336 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qhgbp"] Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.746564 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.770147 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qhgbp"] Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.806645 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-catalog-content\") pod \"redhat-operators-qhgbp\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.806751 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb78j\" (UniqueName: \"kubernetes.io/projected/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-kube-api-access-xb78j\") pod \"redhat-operators-qhgbp\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.806889 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-utilities\") pod \"redhat-operators-qhgbp\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.908365 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-utilities\") pod \"redhat-operators-qhgbp\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.908422 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-catalog-content\") pod \"redhat-operators-qhgbp\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.908500 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb78j\" (UniqueName: \"kubernetes.io/projected/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-kube-api-access-xb78j\") pod \"redhat-operators-qhgbp\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.909008 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-utilities\") pod \"redhat-operators-qhgbp\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.909047 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-catalog-content\") pod \"redhat-operators-qhgbp\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:16 crc kubenswrapper[4575]: I1004 05:32:16.929917 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb78j\" (UniqueName: \"kubernetes.io/projected/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-kube-api-access-xb78j\") pod \"redhat-operators-qhgbp\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:17 crc kubenswrapper[4575]: I1004 05:32:17.069834 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:17 crc kubenswrapper[4575]: I1004 05:32:17.645916 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qhgbp"] Oct 04 05:32:18 crc kubenswrapper[4575]: I1004 05:32:18.557570 4575 generic.go:334] "Generic (PLEG): container finished" podID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerID="1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc" exitCode=0 Oct 04 05:32:18 crc kubenswrapper[4575]: I1004 05:32:18.557645 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhgbp" event={"ID":"562e5aa8-8635-42d5-b48d-b9eb6b24aa61","Type":"ContainerDied","Data":"1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc"} Oct 04 05:32:18 crc kubenswrapper[4575]: I1004 05:32:18.559224 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhgbp" event={"ID":"562e5aa8-8635-42d5-b48d-b9eb6b24aa61","Type":"ContainerStarted","Data":"3b90d178ccdc354c699540e1a99a86985e2a100d202d1cee3ce8c0569114df98"} Oct 04 05:32:20 crc kubenswrapper[4575]: I1004 05:32:20.579883 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhgbp" event={"ID":"562e5aa8-8635-42d5-b48d-b9eb6b24aa61","Type":"ContainerStarted","Data":"fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166"} Oct 04 05:32:24 crc kubenswrapper[4575]: I1004 05:32:24.652538 4575 generic.go:334] "Generic (PLEG): container finished" podID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerID="fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166" exitCode=0 Oct 04 05:32:24 crc kubenswrapper[4575]: I1004 05:32:24.652626 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhgbp" event={"ID":"562e5aa8-8635-42d5-b48d-b9eb6b24aa61","Type":"ContainerDied","Data":"fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166"} Oct 04 05:32:24 crc kubenswrapper[4575]: I1004 05:32:24.786285 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:24 crc kubenswrapper[4575]: I1004 05:32:24.786652 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:25 crc kubenswrapper[4575]: I1004 05:32:25.666517 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhgbp" event={"ID":"562e5aa8-8635-42d5-b48d-b9eb6b24aa61","Type":"ContainerStarted","Data":"00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e"} Oct 04 05:32:25 crc kubenswrapper[4575]: I1004 05:32:25.687728 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qhgbp" podStartSLOduration=2.899964832 podStartE2EDuration="9.687704679s" podCreationTimestamp="2025-10-04 05:32:16 +0000 UTC" firstStartedPulling="2025-10-04 05:32:18.563700213 +0000 UTC m=+3489.892259027" lastFinishedPulling="2025-10-04 05:32:25.35144006 +0000 UTC m=+3496.679998874" observedRunningTime="2025-10-04 05:32:25.685877856 +0000 UTC m=+3497.014436670" watchObservedRunningTime="2025-10-04 05:32:25.687704679 +0000 UTC m=+3497.016263493" Oct 04 05:32:25 crc kubenswrapper[4575]: I1004 05:32:25.845834 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-bcrcn" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerName="registry-server" probeResult="failure" output=< Oct 04 05:32:25 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:32:25 crc kubenswrapper[4575]: > Oct 04 05:32:27 crc kubenswrapper[4575]: I1004 05:32:27.070813 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:27 crc kubenswrapper[4575]: I1004 05:32:27.071239 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:28 crc kubenswrapper[4575]: I1004 05:32:28.140948 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qhgbp" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="registry-server" probeResult="failure" output=< Oct 04 05:32:28 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:32:28 crc kubenswrapper[4575]: > Oct 04 05:32:34 crc kubenswrapper[4575]: I1004 05:32:34.852474 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:34 crc kubenswrapper[4575]: I1004 05:32:34.914870 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:35 crc kubenswrapper[4575]: I1004 05:32:35.655717 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bcrcn"] Oct 04 05:32:36 crc kubenswrapper[4575]: I1004 05:32:36.779972 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bcrcn" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerName="registry-server" containerID="cri-o://9231e4e6ee5c1faab5b44f9131ceedb4c3e6cc5dfc632e43fc03f1d383b7a4f6" gracePeriod=2 Oct 04 05:32:37 crc kubenswrapper[4575]: I1004 05:32:37.802181 4575 generic.go:334] "Generic (PLEG): container finished" podID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerID="9231e4e6ee5c1faab5b44f9131ceedb4c3e6cc5dfc632e43fc03f1d383b7a4f6" exitCode=0 Oct 04 05:32:37 crc kubenswrapper[4575]: I1004 05:32:37.802570 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcrcn" event={"ID":"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb","Type":"ContainerDied","Data":"9231e4e6ee5c1faab5b44f9131ceedb4c3e6cc5dfc632e43fc03f1d383b7a4f6"} Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.123298 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qhgbp" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="registry-server" probeResult="failure" output=< Oct 04 05:32:38 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:32:38 crc kubenswrapper[4575]: > Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.293283 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.414529 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-catalog-content\") pod \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.414873 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-utilities\") pod \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.415070 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhfgn\" (UniqueName: \"kubernetes.io/projected/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-kube-api-access-fhfgn\") pod \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\" (UID: \"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb\") " Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.417141 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-utilities" (OuterVolumeSpecName: "utilities") pod "2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" (UID: "2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.433074 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-kube-api-access-fhfgn" (OuterVolumeSpecName: "kube-api-access-fhfgn") pod "2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" (UID: "2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb"). InnerVolumeSpecName "kube-api-access-fhfgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.446403 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.446652 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.446797 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.447686 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2adda20348c666e32616a3a23aff161d88262e97b92e05c789aa15c41b2f3d1"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.447830 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://c2adda20348c666e32616a3a23aff161d88262e97b92e05c789aa15c41b2f3d1" gracePeriod=600 Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.470862 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" (UID: "2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.517928 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.517964 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.517977 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhfgn\" (UniqueName: \"kubernetes.io/projected/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb-kube-api-access-fhfgn\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.813457 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="c2adda20348c666e32616a3a23aff161d88262e97b92e05c789aa15c41b2f3d1" exitCode=0 Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.813498 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"c2adda20348c666e32616a3a23aff161d88262e97b92e05c789aa15c41b2f3d1"} Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.813922 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087"} Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.813946 4575 scope.go:117] "RemoveContainer" containerID="623cbad2081c29e4e03f6674c1941ff808f95d7ca2e68cc55e6b1bea1aaf0b26" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.818441 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bcrcn" event={"ID":"2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb","Type":"ContainerDied","Data":"9ddac1f0eb40651005f089648eadd7e45a5b536bf9682579b2841ac5d0204759"} Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.818556 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bcrcn" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.856173 4575 scope.go:117] "RemoveContainer" containerID="9231e4e6ee5c1faab5b44f9131ceedb4c3e6cc5dfc632e43fc03f1d383b7a4f6" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.863267 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bcrcn"] Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.875703 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bcrcn"] Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.878081 4575 scope.go:117] "RemoveContainer" containerID="95d28f711e1eef09df34a191ae492eccd899cebda29a737674857db7beeaa9cc" Oct 04 05:32:38 crc kubenswrapper[4575]: I1004 05:32:38.904569 4575 scope.go:117] "RemoveContainer" containerID="4b7fad6d46d7bccf33e5cb34d2e1be81f5d468579ce1fc9fb3564045974b46cf" Oct 04 05:32:39 crc kubenswrapper[4575]: I1004 05:32:39.327501 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" path="/var/lib/kubelet/pods/2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb/volumes" Oct 04 05:32:47 crc kubenswrapper[4575]: I1004 05:32:47.118136 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:47 crc kubenswrapper[4575]: I1004 05:32:47.179577 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:47 crc kubenswrapper[4575]: I1004 05:32:47.948045 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qhgbp"] Oct 04 05:32:48 crc kubenswrapper[4575]: I1004 05:32:48.908912 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qhgbp" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="registry-server" containerID="cri-o://00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e" gracePeriod=2 Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.489968 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.569857 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-catalog-content\") pod \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.570267 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb78j\" (UniqueName: \"kubernetes.io/projected/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-kube-api-access-xb78j\") pod \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.570358 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-utilities\") pod \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\" (UID: \"562e5aa8-8635-42d5-b48d-b9eb6b24aa61\") " Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.577020 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-utilities" (OuterVolumeSpecName: "utilities") pod "562e5aa8-8635-42d5-b48d-b9eb6b24aa61" (UID: "562e5aa8-8635-42d5-b48d-b9eb6b24aa61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.580352 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-kube-api-access-xb78j" (OuterVolumeSpecName: "kube-api-access-xb78j") pod "562e5aa8-8635-42d5-b48d-b9eb6b24aa61" (UID: "562e5aa8-8635-42d5-b48d-b9eb6b24aa61"). InnerVolumeSpecName "kube-api-access-xb78j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.672602 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb78j\" (UniqueName: \"kubernetes.io/projected/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-kube-api-access-xb78j\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.672651 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.678355 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "562e5aa8-8635-42d5-b48d-b9eb6b24aa61" (UID: "562e5aa8-8635-42d5-b48d-b9eb6b24aa61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.775094 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/562e5aa8-8635-42d5-b48d-b9eb6b24aa61-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.926104 4575 generic.go:334] "Generic (PLEG): container finished" podID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerID="00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e" exitCode=0 Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.926196 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhgbp" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.926211 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhgbp" event={"ID":"562e5aa8-8635-42d5-b48d-b9eb6b24aa61","Type":"ContainerDied","Data":"00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e"} Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.927968 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhgbp" event={"ID":"562e5aa8-8635-42d5-b48d-b9eb6b24aa61","Type":"ContainerDied","Data":"3b90d178ccdc354c699540e1a99a86985e2a100d202d1cee3ce8c0569114df98"} Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.928043 4575 scope.go:117] "RemoveContainer" containerID="00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.975714 4575 scope.go:117] "RemoveContainer" containerID="fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166" Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.981480 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qhgbp"] Oct 04 05:32:49 crc kubenswrapper[4575]: I1004 05:32:49.991686 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qhgbp"] Oct 04 05:32:50 crc kubenswrapper[4575]: I1004 05:32:50.000082 4575 scope.go:117] "RemoveContainer" containerID="1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc" Oct 04 05:32:50 crc kubenswrapper[4575]: I1004 05:32:50.045734 4575 scope.go:117] "RemoveContainer" containerID="00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e" Oct 04 05:32:50 crc kubenswrapper[4575]: E1004 05:32:50.046208 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e\": container with ID starting with 00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e not found: ID does not exist" containerID="00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e" Oct 04 05:32:50 crc kubenswrapper[4575]: I1004 05:32:50.046244 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e"} err="failed to get container status \"00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e\": rpc error: code = NotFound desc = could not find container \"00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e\": container with ID starting with 00c080430b6c22eeac4c45d922166daf995a2329d9b0cec173befb86f1b8276e not found: ID does not exist" Oct 04 05:32:50 crc kubenswrapper[4575]: I1004 05:32:50.046271 4575 scope.go:117] "RemoveContainer" containerID="fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166" Oct 04 05:32:50 crc kubenswrapper[4575]: E1004 05:32:50.046560 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166\": container with ID starting with fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166 not found: ID does not exist" containerID="fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166" Oct 04 05:32:50 crc kubenswrapper[4575]: I1004 05:32:50.046620 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166"} err="failed to get container status \"fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166\": rpc error: code = NotFound desc = could not find container \"fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166\": container with ID starting with fa661503807eef5b8c58774c53dcb6cde40b4322ebf6222f419ed32ec1219166 not found: ID does not exist" Oct 04 05:32:50 crc kubenswrapper[4575]: I1004 05:32:50.046636 4575 scope.go:117] "RemoveContainer" containerID="1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc" Oct 04 05:32:50 crc kubenswrapper[4575]: E1004 05:32:50.046905 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc\": container with ID starting with 1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc not found: ID does not exist" containerID="1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc" Oct 04 05:32:50 crc kubenswrapper[4575]: I1004 05:32:50.046926 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc"} err="failed to get container status \"1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc\": rpc error: code = NotFound desc = could not find container \"1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc\": container with ID starting with 1d63d3f39d4c663f067678d39669986c34f99c5faa6f0f993e025075a11d41dc not found: ID does not exist" Oct 04 05:32:51 crc kubenswrapper[4575]: I1004 05:32:51.324025 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" path="/var/lib/kubelet/pods/562e5aa8-8635-42d5-b48d-b9eb6b24aa61/volumes" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.685001 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ddqtk"] Oct 04 05:33:18 crc kubenswrapper[4575]: E1004 05:33:18.688238 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="extract-utilities" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.688272 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="extract-utilities" Oct 04 05:33:18 crc kubenswrapper[4575]: E1004 05:33:18.688288 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerName="registry-server" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.688296 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerName="registry-server" Oct 04 05:33:18 crc kubenswrapper[4575]: E1004 05:33:18.688321 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerName="extract-utilities" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.688330 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerName="extract-utilities" Oct 04 05:33:18 crc kubenswrapper[4575]: E1004 05:33:18.688351 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerName="extract-content" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.688359 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerName="extract-content" Oct 04 05:33:18 crc kubenswrapper[4575]: E1004 05:33:18.688390 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="extract-content" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.688400 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="extract-content" Oct 04 05:33:18 crc kubenswrapper[4575]: E1004 05:33:18.688418 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="registry-server" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.688428 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="registry-server" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.688714 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b4e00cc-e80d-4b4e-a52f-40c56b76b6eb" containerName="registry-server" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.688741 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="562e5aa8-8635-42d5-b48d-b9eb6b24aa61" containerName="registry-server" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.693678 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.717524 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddqtk"] Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.795839 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmxzc\" (UniqueName: \"kubernetes.io/projected/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-kube-api-access-qmxzc\") pod \"redhat-marketplace-ddqtk\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.796226 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-utilities\") pod \"redhat-marketplace-ddqtk\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.796449 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-catalog-content\") pod \"redhat-marketplace-ddqtk\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.898544 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmxzc\" (UniqueName: \"kubernetes.io/projected/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-kube-api-access-qmxzc\") pod \"redhat-marketplace-ddqtk\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.898701 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-utilities\") pod \"redhat-marketplace-ddqtk\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.898823 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-catalog-content\") pod \"redhat-marketplace-ddqtk\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.899425 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-catalog-content\") pod \"redhat-marketplace-ddqtk\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.899779 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-utilities\") pod \"redhat-marketplace-ddqtk\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:18 crc kubenswrapper[4575]: I1004 05:33:18.926609 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmxzc\" (UniqueName: \"kubernetes.io/projected/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-kube-api-access-qmxzc\") pod \"redhat-marketplace-ddqtk\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:19 crc kubenswrapper[4575]: I1004 05:33:19.022861 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:19 crc kubenswrapper[4575]: I1004 05:33:19.710339 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddqtk"] Oct 04 05:33:20 crc kubenswrapper[4575]: I1004 05:33:20.205752 4575 generic.go:334] "Generic (PLEG): container finished" podID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerID="d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b" exitCode=0 Oct 04 05:33:20 crc kubenswrapper[4575]: I1004 05:33:20.205885 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddqtk" event={"ID":"b529b41f-569d-47d0-ae8c-8fa426e8d0f8","Type":"ContainerDied","Data":"d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b"} Oct 04 05:33:20 crc kubenswrapper[4575]: I1004 05:33:20.206216 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddqtk" event={"ID":"b529b41f-569d-47d0-ae8c-8fa426e8d0f8","Type":"ContainerStarted","Data":"713571d941e3a6cad5cdb9647f3f379fe9052440d1a969ca7ff6fd7c1e53791e"} Oct 04 05:33:21 crc kubenswrapper[4575]: I1004 05:33:21.219011 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddqtk" event={"ID":"b529b41f-569d-47d0-ae8c-8fa426e8d0f8","Type":"ContainerStarted","Data":"964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d"} Oct 04 05:33:22 crc kubenswrapper[4575]: I1004 05:33:22.230285 4575 generic.go:334] "Generic (PLEG): container finished" podID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerID="964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d" exitCode=0 Oct 04 05:33:22 crc kubenswrapper[4575]: I1004 05:33:22.230790 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddqtk" event={"ID":"b529b41f-569d-47d0-ae8c-8fa426e8d0f8","Type":"ContainerDied","Data":"964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d"} Oct 04 05:33:23 crc kubenswrapper[4575]: I1004 05:33:23.248576 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddqtk" event={"ID":"b529b41f-569d-47d0-ae8c-8fa426e8d0f8","Type":"ContainerStarted","Data":"0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b"} Oct 04 05:33:23 crc kubenswrapper[4575]: I1004 05:33:23.281276 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ddqtk" podStartSLOduration=2.822453939 podStartE2EDuration="5.281251463s" podCreationTimestamp="2025-10-04 05:33:18 +0000 UTC" firstStartedPulling="2025-10-04 05:33:20.208970414 +0000 UTC m=+3551.537529228" lastFinishedPulling="2025-10-04 05:33:22.667767938 +0000 UTC m=+3553.996326752" observedRunningTime="2025-10-04 05:33:23.269101973 +0000 UTC m=+3554.597660787" watchObservedRunningTime="2025-10-04 05:33:23.281251463 +0000 UTC m=+3554.609810277" Oct 04 05:33:29 crc kubenswrapper[4575]: I1004 05:33:29.023439 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:29 crc kubenswrapper[4575]: I1004 05:33:29.023884 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:29 crc kubenswrapper[4575]: I1004 05:33:29.070836 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:29 crc kubenswrapper[4575]: I1004 05:33:29.380274 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:29 crc kubenswrapper[4575]: I1004 05:33:29.434072 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddqtk"] Oct 04 05:33:31 crc kubenswrapper[4575]: I1004 05:33:31.345027 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ddqtk" podUID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerName="registry-server" containerID="cri-o://0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b" gracePeriod=2 Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.045816 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.222138 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-catalog-content\") pod \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.222190 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-utilities\") pod \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.222438 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmxzc\" (UniqueName: \"kubernetes.io/projected/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-kube-api-access-qmxzc\") pod \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\" (UID: \"b529b41f-569d-47d0-ae8c-8fa426e8d0f8\") " Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.223433 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-utilities" (OuterVolumeSpecName: "utilities") pod "b529b41f-569d-47d0-ae8c-8fa426e8d0f8" (UID: "b529b41f-569d-47d0-ae8c-8fa426e8d0f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.235532 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-kube-api-access-qmxzc" (OuterVolumeSpecName: "kube-api-access-qmxzc") pod "b529b41f-569d-47d0-ae8c-8fa426e8d0f8" (UID: "b529b41f-569d-47d0-ae8c-8fa426e8d0f8"). InnerVolumeSpecName "kube-api-access-qmxzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.244720 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b529b41f-569d-47d0-ae8c-8fa426e8d0f8" (UID: "b529b41f-569d-47d0-ae8c-8fa426e8d0f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.325726 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmxzc\" (UniqueName: \"kubernetes.io/projected/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-kube-api-access-qmxzc\") on node \"crc\" DevicePath \"\"" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.325777 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.325787 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529b41f-569d-47d0-ae8c-8fa426e8d0f8-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.357925 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddqtk" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.357966 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddqtk" event={"ID":"b529b41f-569d-47d0-ae8c-8fa426e8d0f8","Type":"ContainerDied","Data":"0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b"} Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.358121 4575 scope.go:117] "RemoveContainer" containerID="0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.357568 4575 generic.go:334] "Generic (PLEG): container finished" podID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerID="0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b" exitCode=0 Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.358258 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddqtk" event={"ID":"b529b41f-569d-47d0-ae8c-8fa426e8d0f8","Type":"ContainerDied","Data":"713571d941e3a6cad5cdb9647f3f379fe9052440d1a969ca7ff6fd7c1e53791e"} Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.388139 4575 scope.go:117] "RemoveContainer" containerID="964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.412883 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddqtk"] Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.424190 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddqtk"] Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.442468 4575 scope.go:117] "RemoveContainer" containerID="d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.497715 4575 scope.go:117] "RemoveContainer" containerID="0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b" Oct 04 05:33:32 crc kubenswrapper[4575]: E1004 05:33:32.500084 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b\": container with ID starting with 0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b not found: ID does not exist" containerID="0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.500142 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b"} err="failed to get container status \"0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b\": rpc error: code = NotFound desc = could not find container \"0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b\": container with ID starting with 0cb9925bb3c66117dc195c52fd9b41516e2940feac2da2d8d445a0421564e42b not found: ID does not exist" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.500172 4575 scope.go:117] "RemoveContainer" containerID="964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d" Oct 04 05:33:32 crc kubenswrapper[4575]: E1004 05:33:32.500571 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d\": container with ID starting with 964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d not found: ID does not exist" containerID="964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.500627 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d"} err="failed to get container status \"964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d\": rpc error: code = NotFound desc = could not find container \"964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d\": container with ID starting with 964d07520c244f5bd638535624408a147c3736e3dabbed9e9ef10eb3c4891e0d not found: ID does not exist" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.500647 4575 scope.go:117] "RemoveContainer" containerID="d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b" Oct 04 05:33:32 crc kubenswrapper[4575]: E1004 05:33:32.501002 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b\": container with ID starting with d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b not found: ID does not exist" containerID="d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b" Oct 04 05:33:32 crc kubenswrapper[4575]: I1004 05:33:32.501037 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b"} err="failed to get container status \"d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b\": rpc error: code = NotFound desc = could not find container \"d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b\": container with ID starting with d014dc32ad5f128c3692bbe7cff3a5cce09b4dbe9d2fbcd3b71b897c8bdd6f8b not found: ID does not exist" Oct 04 05:33:33 crc kubenswrapper[4575]: I1004 05:33:33.321146 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" path="/var/lib/kubelet/pods/b529b41f-569d-47d0-ae8c-8fa426e8d0f8/volumes" Oct 04 05:34:38 crc kubenswrapper[4575]: I1004 05:34:38.446226 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:34:38 crc kubenswrapper[4575]: I1004 05:34:38.446921 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:35:08 crc kubenswrapper[4575]: I1004 05:35:08.445928 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:35:08 crc kubenswrapper[4575]: I1004 05:35:08.446493 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:35:38 crc kubenswrapper[4575]: I1004 05:35:38.445966 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:35:38 crc kubenswrapper[4575]: I1004 05:35:38.446563 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:35:38 crc kubenswrapper[4575]: I1004 05:35:38.446624 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:35:38 crc kubenswrapper[4575]: I1004 05:35:38.460854 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:35:38 crc kubenswrapper[4575]: I1004 05:35:38.461376 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" gracePeriod=600 Oct 04 05:35:38 crc kubenswrapper[4575]: E1004 05:35:38.614101 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:35:39 crc kubenswrapper[4575]: I1004 05:35:39.560405 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" exitCode=0 Oct 04 05:35:39 crc kubenswrapper[4575]: I1004 05:35:39.560454 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087"} Oct 04 05:35:39 crc kubenswrapper[4575]: I1004 05:35:39.560501 4575 scope.go:117] "RemoveContainer" containerID="c2adda20348c666e32616a3a23aff161d88262e97b92e05c789aa15c41b2f3d1" Oct 04 05:35:39 crc kubenswrapper[4575]: I1004 05:35:39.561308 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:35:39 crc kubenswrapper[4575]: E1004 05:35:39.561656 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:35:51 crc kubenswrapper[4575]: I1004 05:35:51.310927 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:35:51 crc kubenswrapper[4575]: E1004 05:35:51.311828 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:36:06 crc kubenswrapper[4575]: I1004 05:36:06.310481 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:36:06 crc kubenswrapper[4575]: E1004 05:36:06.311462 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:36:21 crc kubenswrapper[4575]: I1004 05:36:21.310300 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:36:21 crc kubenswrapper[4575]: E1004 05:36:21.311091 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:36:36 crc kubenswrapper[4575]: I1004 05:36:36.310447 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:36:36 crc kubenswrapper[4575]: E1004 05:36:36.311436 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:36:49 crc kubenswrapper[4575]: I1004 05:36:49.329615 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:36:49 crc kubenswrapper[4575]: E1004 05:36:49.330853 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:37:01 crc kubenswrapper[4575]: I1004 05:37:01.311551 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:37:01 crc kubenswrapper[4575]: E1004 05:37:01.314241 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:37:14 crc kubenswrapper[4575]: I1004 05:37:14.314273 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:37:14 crc kubenswrapper[4575]: E1004 05:37:14.315321 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.848414 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-69k9s"] Oct 04 05:37:22 crc kubenswrapper[4575]: E1004 05:37:22.849408 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerName="registry-server" Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.849422 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerName="registry-server" Oct 04 05:37:22 crc kubenswrapper[4575]: E1004 05:37:22.849454 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerName="extract-utilities" Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.849461 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerName="extract-utilities" Oct 04 05:37:22 crc kubenswrapper[4575]: E1004 05:37:22.849477 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerName="extract-content" Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.849485 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerName="extract-content" Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.849707 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="b529b41f-569d-47d0-ae8c-8fa426e8d0f8" containerName="registry-server" Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.851092 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.868376 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-69k9s"] Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.963417 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-catalog-content\") pod \"community-operators-69k9s\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.963515 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kffn6\" (UniqueName: \"kubernetes.io/projected/4ffee271-22e8-40c0-84c4-e8c2889ae391-kube-api-access-kffn6\") pod \"community-operators-69k9s\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:22 crc kubenswrapper[4575]: I1004 05:37:22.963561 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-utilities\") pod \"community-operators-69k9s\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:23 crc kubenswrapper[4575]: I1004 05:37:23.065069 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-utilities\") pod \"community-operators-69k9s\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:23 crc kubenswrapper[4575]: I1004 05:37:23.065260 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-catalog-content\") pod \"community-operators-69k9s\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:23 crc kubenswrapper[4575]: I1004 05:37:23.065326 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kffn6\" (UniqueName: \"kubernetes.io/projected/4ffee271-22e8-40c0-84c4-e8c2889ae391-kube-api-access-kffn6\") pod \"community-operators-69k9s\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:23 crc kubenswrapper[4575]: I1004 05:37:23.066197 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-utilities\") pod \"community-operators-69k9s\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:23 crc kubenswrapper[4575]: I1004 05:37:23.066482 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-catalog-content\") pod \"community-operators-69k9s\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:23 crc kubenswrapper[4575]: I1004 05:37:23.092645 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kffn6\" (UniqueName: \"kubernetes.io/projected/4ffee271-22e8-40c0-84c4-e8c2889ae391-kube-api-access-kffn6\") pod \"community-operators-69k9s\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:23 crc kubenswrapper[4575]: I1004 05:37:23.179469 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:23 crc kubenswrapper[4575]: I1004 05:37:23.752158 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-69k9s"] Oct 04 05:37:24 crc kubenswrapper[4575]: I1004 05:37:24.621782 4575 generic.go:334] "Generic (PLEG): container finished" podID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerID="94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85" exitCode=0 Oct 04 05:37:24 crc kubenswrapper[4575]: I1004 05:37:24.621860 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69k9s" event={"ID":"4ffee271-22e8-40c0-84c4-e8c2889ae391","Type":"ContainerDied","Data":"94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85"} Oct 04 05:37:24 crc kubenswrapper[4575]: I1004 05:37:24.623779 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69k9s" event={"ID":"4ffee271-22e8-40c0-84c4-e8c2889ae391","Type":"ContainerStarted","Data":"665c403306c1211402ff90020b858bf7d931b0788bc65b6fefbe1517051aaad6"} Oct 04 05:37:24 crc kubenswrapper[4575]: I1004 05:37:24.623997 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:37:25 crc kubenswrapper[4575]: I1004 05:37:25.310429 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:37:25 crc kubenswrapper[4575]: E1004 05:37:25.310711 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:37:25 crc kubenswrapper[4575]: I1004 05:37:25.633321 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69k9s" event={"ID":"4ffee271-22e8-40c0-84c4-e8c2889ae391","Type":"ContainerStarted","Data":"ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172"} Oct 04 05:37:27 crc kubenswrapper[4575]: I1004 05:37:27.654648 4575 generic.go:334] "Generic (PLEG): container finished" podID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerID="ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172" exitCode=0 Oct 04 05:37:27 crc kubenswrapper[4575]: I1004 05:37:27.654764 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69k9s" event={"ID":"4ffee271-22e8-40c0-84c4-e8c2889ae391","Type":"ContainerDied","Data":"ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172"} Oct 04 05:37:29 crc kubenswrapper[4575]: I1004 05:37:29.673853 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69k9s" event={"ID":"4ffee271-22e8-40c0-84c4-e8c2889ae391","Type":"ContainerStarted","Data":"8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48"} Oct 04 05:37:29 crc kubenswrapper[4575]: I1004 05:37:29.701744 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-69k9s" podStartSLOduration=3.969277617 podStartE2EDuration="7.701722277s" podCreationTimestamp="2025-10-04 05:37:22 +0000 UTC" firstStartedPulling="2025-10-04 05:37:24.623704687 +0000 UTC m=+3795.952263501" lastFinishedPulling="2025-10-04 05:37:28.356149347 +0000 UTC m=+3799.684708161" observedRunningTime="2025-10-04 05:37:29.69594461 +0000 UTC m=+3801.024503444" watchObservedRunningTime="2025-10-04 05:37:29.701722277 +0000 UTC m=+3801.030281101" Oct 04 05:37:33 crc kubenswrapper[4575]: I1004 05:37:33.179748 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:33 crc kubenswrapper[4575]: I1004 05:37:33.180798 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:33 crc kubenswrapper[4575]: I1004 05:37:33.255713 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:33 crc kubenswrapper[4575]: I1004 05:37:33.757163 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:33 crc kubenswrapper[4575]: I1004 05:37:33.811668 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-69k9s"] Oct 04 05:37:35 crc kubenswrapper[4575]: I1004 05:37:35.724686 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-69k9s" podUID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerName="registry-server" containerID="cri-o://8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48" gracePeriod=2 Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.413125 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.585709 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kffn6\" (UniqueName: \"kubernetes.io/projected/4ffee271-22e8-40c0-84c4-e8c2889ae391-kube-api-access-kffn6\") pod \"4ffee271-22e8-40c0-84c4-e8c2889ae391\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.585788 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-catalog-content\") pod \"4ffee271-22e8-40c0-84c4-e8c2889ae391\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.585982 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-utilities\") pod \"4ffee271-22e8-40c0-84c4-e8c2889ae391\" (UID: \"4ffee271-22e8-40c0-84c4-e8c2889ae391\") " Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.587446 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-utilities" (OuterVolumeSpecName: "utilities") pod "4ffee271-22e8-40c0-84c4-e8c2889ae391" (UID: "4ffee271-22e8-40c0-84c4-e8c2889ae391"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.594866 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ffee271-22e8-40c0-84c4-e8c2889ae391-kube-api-access-kffn6" (OuterVolumeSpecName: "kube-api-access-kffn6") pod "4ffee271-22e8-40c0-84c4-e8c2889ae391" (UID: "4ffee271-22e8-40c0-84c4-e8c2889ae391"). InnerVolumeSpecName "kube-api-access-kffn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.680367 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ffee271-22e8-40c0-84c4-e8c2889ae391" (UID: "4ffee271-22e8-40c0-84c4-e8c2889ae391"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.688715 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.688756 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kffn6\" (UniqueName: \"kubernetes.io/projected/4ffee271-22e8-40c0-84c4-e8c2889ae391-kube-api-access-kffn6\") on node \"crc\" DevicePath \"\"" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.688767 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ffee271-22e8-40c0-84c4-e8c2889ae391-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.739116 4575 generic.go:334] "Generic (PLEG): container finished" podID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerID="8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48" exitCode=0 Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.739160 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69k9s" event={"ID":"4ffee271-22e8-40c0-84c4-e8c2889ae391","Type":"ContainerDied","Data":"8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48"} Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.739189 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69k9s" event={"ID":"4ffee271-22e8-40c0-84c4-e8c2889ae391","Type":"ContainerDied","Data":"665c403306c1211402ff90020b858bf7d931b0788bc65b6fefbe1517051aaad6"} Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.739209 4575 scope.go:117] "RemoveContainer" containerID="8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.739214 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69k9s" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.778066 4575 scope.go:117] "RemoveContainer" containerID="ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.792966 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-69k9s"] Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.817148 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-69k9s"] Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.843836 4575 scope.go:117] "RemoveContainer" containerID="94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.913220 4575 scope.go:117] "RemoveContainer" containerID="8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48" Oct 04 05:37:36 crc kubenswrapper[4575]: E1004 05:37:36.913871 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48\": container with ID starting with 8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48 not found: ID does not exist" containerID="8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.913902 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48"} err="failed to get container status \"8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48\": rpc error: code = NotFound desc = could not find container \"8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48\": container with ID starting with 8fab102fc5e0a57ca25fd09194b937ccfe582a41629d69972d634fda36e4dd48 not found: ID does not exist" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.913925 4575 scope.go:117] "RemoveContainer" containerID="ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172" Oct 04 05:37:36 crc kubenswrapper[4575]: E1004 05:37:36.914187 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172\": container with ID starting with ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172 not found: ID does not exist" containerID="ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.914212 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172"} err="failed to get container status \"ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172\": rpc error: code = NotFound desc = could not find container \"ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172\": container with ID starting with ad18ddb61516c6613ccee421781164d8ef1968f803e8b74c4331f8354e84f172 not found: ID does not exist" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.914228 4575 scope.go:117] "RemoveContainer" containerID="94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85" Oct 04 05:37:36 crc kubenswrapper[4575]: E1004 05:37:36.914417 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85\": container with ID starting with 94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85 not found: ID does not exist" containerID="94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85" Oct 04 05:37:36 crc kubenswrapper[4575]: I1004 05:37:36.914440 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85"} err="failed to get container status \"94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85\": rpc error: code = NotFound desc = could not find container \"94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85\": container with ID starting with 94cb5757440c2ef2ed766b4465e154b0752ff8ac1e2cec25999b059a5a36ea85 not found: ID does not exist" Oct 04 05:37:37 crc kubenswrapper[4575]: I1004 05:37:37.311038 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:37:37 crc kubenswrapper[4575]: E1004 05:37:37.311623 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:37:37 crc kubenswrapper[4575]: I1004 05:37:37.322894 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ffee271-22e8-40c0-84c4-e8c2889ae391" path="/var/lib/kubelet/pods/4ffee271-22e8-40c0-84c4-e8c2889ae391/volumes" Oct 04 05:37:51 crc kubenswrapper[4575]: I1004 05:37:51.310441 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:37:51 crc kubenswrapper[4575]: E1004 05:37:51.311307 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:38:05 crc kubenswrapper[4575]: I1004 05:38:05.309877 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:38:05 crc kubenswrapper[4575]: E1004 05:38:05.310691 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:38:16 crc kubenswrapper[4575]: I1004 05:38:16.310817 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:38:16 crc kubenswrapper[4575]: E1004 05:38:16.311552 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:38:27 crc kubenswrapper[4575]: I1004 05:38:27.310579 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:38:27 crc kubenswrapper[4575]: E1004 05:38:27.312023 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:38:38 crc kubenswrapper[4575]: I1004 05:38:38.309980 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:38:38 crc kubenswrapper[4575]: E1004 05:38:38.310864 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:38:51 crc kubenswrapper[4575]: I1004 05:38:51.310769 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:38:51 crc kubenswrapper[4575]: E1004 05:38:51.311895 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:39:02 crc kubenswrapper[4575]: I1004 05:39:02.309941 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:39:02 crc kubenswrapper[4575]: E1004 05:39:02.311051 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:39:14 crc kubenswrapper[4575]: I1004 05:39:14.310062 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:39:14 crc kubenswrapper[4575]: E1004 05:39:14.311173 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:39:27 crc kubenswrapper[4575]: I1004 05:39:27.311679 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:39:27 crc kubenswrapper[4575]: E1004 05:39:27.312782 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:39:40 crc kubenswrapper[4575]: I1004 05:39:40.310554 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:39:40 crc kubenswrapper[4575]: E1004 05:39:40.311547 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:39:54 crc kubenswrapper[4575]: I1004 05:39:54.310209 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:39:54 crc kubenswrapper[4575]: E1004 05:39:54.311396 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:40:06 crc kubenswrapper[4575]: I1004 05:40:06.310482 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:40:06 crc kubenswrapper[4575]: E1004 05:40:06.311219 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:40:20 crc kubenswrapper[4575]: I1004 05:40:20.310297 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:40:20 crc kubenswrapper[4575]: E1004 05:40:20.311170 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:40:33 crc kubenswrapper[4575]: I1004 05:40:33.311925 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:40:33 crc kubenswrapper[4575]: E1004 05:40:33.312968 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:40:48 crc kubenswrapper[4575]: I1004 05:40:48.311104 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:40:48 crc kubenswrapper[4575]: I1004 05:40:48.785363 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"8dce56ed3e369405293833fa9a23c605f7ddc95e08b651fd4ef5764d7d1df662"} Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.650016 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ltfxf"] Oct 04 05:42:30 crc kubenswrapper[4575]: E1004 05:42:30.651310 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerName="registry-server" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.651328 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerName="registry-server" Oct 04 05:42:30 crc kubenswrapper[4575]: E1004 05:42:30.651365 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerName="extract-utilities" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.651371 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerName="extract-utilities" Oct 04 05:42:30 crc kubenswrapper[4575]: E1004 05:42:30.651389 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerName="extract-content" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.651396 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerName="extract-content" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.651795 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ffee271-22e8-40c0-84c4-e8c2889ae391" containerName="registry-server" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.653254 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.676379 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ltfxf"] Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.761512 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-utilities\") pod \"certified-operators-ltfxf\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.762035 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-catalog-content\") pod \"certified-operators-ltfxf\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.762183 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4njnp\" (UniqueName: \"kubernetes.io/projected/436e1da4-8740-48d1-8823-28d8dfa7715f-kube-api-access-4njnp\") pod \"certified-operators-ltfxf\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.865070 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-catalog-content\") pod \"certified-operators-ltfxf\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.865206 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4njnp\" (UniqueName: \"kubernetes.io/projected/436e1da4-8740-48d1-8823-28d8dfa7715f-kube-api-access-4njnp\") pod \"certified-operators-ltfxf\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.865347 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-utilities\") pod \"certified-operators-ltfxf\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.865952 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-catalog-content\") pod \"certified-operators-ltfxf\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.865965 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-utilities\") pod \"certified-operators-ltfxf\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.891855 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4njnp\" (UniqueName: \"kubernetes.io/projected/436e1da4-8740-48d1-8823-28d8dfa7715f-kube-api-access-4njnp\") pod \"certified-operators-ltfxf\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:30 crc kubenswrapper[4575]: I1004 05:42:30.989223 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:31 crc kubenswrapper[4575]: I1004 05:42:31.663511 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ltfxf"] Oct 04 05:42:31 crc kubenswrapper[4575]: I1004 05:42:31.865536 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltfxf" event={"ID":"436e1da4-8740-48d1-8823-28d8dfa7715f","Type":"ContainerStarted","Data":"3ab04b7546de8508c8447aa657800a455c64c6e7f87752c9a57a772f349e5e68"} Oct 04 05:42:32 crc kubenswrapper[4575]: I1004 05:42:32.876388 4575 generic.go:334] "Generic (PLEG): container finished" podID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerID="d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551" exitCode=0 Oct 04 05:42:32 crc kubenswrapper[4575]: I1004 05:42:32.876806 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltfxf" event={"ID":"436e1da4-8740-48d1-8823-28d8dfa7715f","Type":"ContainerDied","Data":"d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551"} Oct 04 05:42:32 crc kubenswrapper[4575]: I1004 05:42:32.879026 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:42:34 crc kubenswrapper[4575]: I1004 05:42:34.899003 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltfxf" event={"ID":"436e1da4-8740-48d1-8823-28d8dfa7715f","Type":"ContainerStarted","Data":"89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594"} Oct 04 05:42:35 crc kubenswrapper[4575]: I1004 05:42:35.911955 4575 generic.go:334] "Generic (PLEG): container finished" podID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerID="89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594" exitCode=0 Oct 04 05:42:35 crc kubenswrapper[4575]: I1004 05:42:35.912052 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltfxf" event={"ID":"436e1da4-8740-48d1-8823-28d8dfa7715f","Type":"ContainerDied","Data":"89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594"} Oct 04 05:42:36 crc kubenswrapper[4575]: I1004 05:42:36.928201 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltfxf" event={"ID":"436e1da4-8740-48d1-8823-28d8dfa7715f","Type":"ContainerStarted","Data":"e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d"} Oct 04 05:42:36 crc kubenswrapper[4575]: I1004 05:42:36.953028 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ltfxf" podStartSLOduration=3.517264657 podStartE2EDuration="6.952999774s" podCreationTimestamp="2025-10-04 05:42:30 +0000 UTC" firstStartedPulling="2025-10-04 05:42:32.878818056 +0000 UTC m=+4104.207376870" lastFinishedPulling="2025-10-04 05:42:36.314553173 +0000 UTC m=+4107.643111987" observedRunningTime="2025-10-04 05:42:36.949104011 +0000 UTC m=+4108.277662855" watchObservedRunningTime="2025-10-04 05:42:36.952999774 +0000 UTC m=+4108.281558598" Oct 04 05:42:40 crc kubenswrapper[4575]: I1004 05:42:40.991251 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:40 crc kubenswrapper[4575]: I1004 05:42:40.991905 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:41 crc kubenswrapper[4575]: I1004 05:42:41.044524 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:42 crc kubenswrapper[4575]: I1004 05:42:42.033916 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:42 crc kubenswrapper[4575]: I1004 05:42:42.093542 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ltfxf"] Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.001371 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ltfxf" podUID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerName="registry-server" containerID="cri-o://e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d" gracePeriod=2 Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.681525 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.782276 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-utilities\") pod \"436e1da4-8740-48d1-8823-28d8dfa7715f\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.782350 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-catalog-content\") pod \"436e1da4-8740-48d1-8823-28d8dfa7715f\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.782659 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4njnp\" (UniqueName: \"kubernetes.io/projected/436e1da4-8740-48d1-8823-28d8dfa7715f-kube-api-access-4njnp\") pod \"436e1da4-8740-48d1-8823-28d8dfa7715f\" (UID: \"436e1da4-8740-48d1-8823-28d8dfa7715f\") " Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.783889 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-utilities" (OuterVolumeSpecName: "utilities") pod "436e1da4-8740-48d1-8823-28d8dfa7715f" (UID: "436e1da4-8740-48d1-8823-28d8dfa7715f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.793542 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/436e1da4-8740-48d1-8823-28d8dfa7715f-kube-api-access-4njnp" (OuterVolumeSpecName: "kube-api-access-4njnp") pod "436e1da4-8740-48d1-8823-28d8dfa7715f" (UID: "436e1da4-8740-48d1-8823-28d8dfa7715f"). InnerVolumeSpecName "kube-api-access-4njnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.886085 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.886133 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4njnp\" (UniqueName: \"kubernetes.io/projected/436e1da4-8740-48d1-8823-28d8dfa7715f-kube-api-access-4njnp\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.920765 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "436e1da4-8740-48d1-8823-28d8dfa7715f" (UID: "436e1da4-8740-48d1-8823-28d8dfa7715f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:42:44 crc kubenswrapper[4575]: I1004 05:42:44.987693 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436e1da4-8740-48d1-8823-28d8dfa7715f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.011990 4575 generic.go:334] "Generic (PLEG): container finished" podID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerID="e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d" exitCode=0 Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.012042 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltfxf" event={"ID":"436e1da4-8740-48d1-8823-28d8dfa7715f","Type":"ContainerDied","Data":"e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d"} Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.012355 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltfxf" event={"ID":"436e1da4-8740-48d1-8823-28d8dfa7715f","Type":"ContainerDied","Data":"3ab04b7546de8508c8447aa657800a455c64c6e7f87752c9a57a772f349e5e68"} Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.012379 4575 scope.go:117] "RemoveContainer" containerID="e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.012562 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltfxf" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.061494 4575 scope.go:117] "RemoveContainer" containerID="89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.065092 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ltfxf"] Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.083140 4575 scope.go:117] "RemoveContainer" containerID="d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.085185 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ltfxf"] Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.130038 4575 scope.go:117] "RemoveContainer" containerID="e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d" Oct 04 05:42:45 crc kubenswrapper[4575]: E1004 05:42:45.130550 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d\": container with ID starting with e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d not found: ID does not exist" containerID="e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.130619 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d"} err="failed to get container status \"e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d\": rpc error: code = NotFound desc = could not find container \"e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d\": container with ID starting with e6fdc8d1e5dcd06fe8c0d501774320e2d236b69fdba41c20e83279350520d84d not found: ID does not exist" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.130655 4575 scope.go:117] "RemoveContainer" containerID="89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594" Oct 04 05:42:45 crc kubenswrapper[4575]: E1004 05:42:45.131076 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594\": container with ID starting with 89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594 not found: ID does not exist" containerID="89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.131217 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594"} err="failed to get container status \"89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594\": rpc error: code = NotFound desc = could not find container \"89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594\": container with ID starting with 89cce22fcf7f6f3ae9ad3945086abfa10da7c93b1df402f76ead784549a2c594 not found: ID does not exist" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.131256 4575 scope.go:117] "RemoveContainer" containerID="d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551" Oct 04 05:42:45 crc kubenswrapper[4575]: E1004 05:42:45.131513 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551\": container with ID starting with d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551 not found: ID does not exist" containerID="d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.131539 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551"} err="failed to get container status \"d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551\": rpc error: code = NotFound desc = could not find container \"d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551\": container with ID starting with d46a3052f8ec791d67f93a4ec922b850360b0219520d44e99cf7ce7548739551 not found: ID does not exist" Oct 04 05:42:45 crc kubenswrapper[4575]: I1004 05:42:45.322331 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="436e1da4-8740-48d1-8823-28d8dfa7715f" path="/var/lib/kubelet/pods/436e1da4-8740-48d1-8823-28d8dfa7715f/volumes" Oct 04 05:43:08 crc kubenswrapper[4575]: I1004 05:43:08.445821 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:43:08 crc kubenswrapper[4575]: I1004 05:43:08.447030 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:43:38 crc kubenswrapper[4575]: I1004 05:43:38.446840 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:43:38 crc kubenswrapper[4575]: I1004 05:43:38.447494 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.805285 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nj28v"] Oct 04 05:43:44 crc kubenswrapper[4575]: E1004 05:43:44.806483 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerName="extract-content" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.806503 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerName="extract-content" Oct 04 05:43:44 crc kubenswrapper[4575]: E1004 05:43:44.806517 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerName="registry-server" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.806525 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerName="registry-server" Oct 04 05:43:44 crc kubenswrapper[4575]: E1004 05:43:44.806539 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerName="extract-utilities" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.806547 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerName="extract-utilities" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.806795 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="436e1da4-8740-48d1-8823-28d8dfa7715f" containerName="registry-server" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.808654 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.829266 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj28v"] Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.860899 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-catalog-content\") pod \"redhat-marketplace-nj28v\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.861071 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rlg4\" (UniqueName: \"kubernetes.io/projected/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-kube-api-access-5rlg4\") pod \"redhat-marketplace-nj28v\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.861106 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-utilities\") pod \"redhat-marketplace-nj28v\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.961951 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rlg4\" (UniqueName: \"kubernetes.io/projected/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-kube-api-access-5rlg4\") pod \"redhat-marketplace-nj28v\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.962028 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-utilities\") pod \"redhat-marketplace-nj28v\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.962066 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-catalog-content\") pod \"redhat-marketplace-nj28v\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.962730 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-catalog-content\") pod \"redhat-marketplace-nj28v\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:44 crc kubenswrapper[4575]: I1004 05:43:44.963046 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-utilities\") pod \"redhat-marketplace-nj28v\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:45 crc kubenswrapper[4575]: I1004 05:43:45.002169 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rlg4\" (UniqueName: \"kubernetes.io/projected/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-kube-api-access-5rlg4\") pod \"redhat-marketplace-nj28v\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:45 crc kubenswrapper[4575]: I1004 05:43:45.140381 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:45 crc kubenswrapper[4575]: I1004 05:43:45.910524 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj28v"] Oct 04 05:43:46 crc kubenswrapper[4575]: I1004 05:43:46.651706 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj28v" event={"ID":"7f1c3aac-6b1f-48e4-846b-192e0977fb7a","Type":"ContainerStarted","Data":"ac6d7132c364bc0761d549247a0412e1e402012a64e256ed21456e8fb860712e"} Oct 04 05:43:47 crc kubenswrapper[4575]: I1004 05:43:47.664875 4575 generic.go:334] "Generic (PLEG): container finished" podID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerID="7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d" exitCode=0 Oct 04 05:43:47 crc kubenswrapper[4575]: I1004 05:43:47.664973 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj28v" event={"ID":"7f1c3aac-6b1f-48e4-846b-192e0977fb7a","Type":"ContainerDied","Data":"7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d"} Oct 04 05:43:49 crc kubenswrapper[4575]: I1004 05:43:49.687564 4575 generic.go:334] "Generic (PLEG): container finished" podID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerID="4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488" exitCode=0 Oct 04 05:43:49 crc kubenswrapper[4575]: I1004 05:43:49.687688 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj28v" event={"ID":"7f1c3aac-6b1f-48e4-846b-192e0977fb7a","Type":"ContainerDied","Data":"4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488"} Oct 04 05:43:50 crc kubenswrapper[4575]: I1004 05:43:50.702388 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj28v" event={"ID":"7f1c3aac-6b1f-48e4-846b-192e0977fb7a","Type":"ContainerStarted","Data":"1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf"} Oct 04 05:43:50 crc kubenswrapper[4575]: I1004 05:43:50.760644 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nj28v" podStartSLOduration=4.247620981 podStartE2EDuration="6.760623662s" podCreationTimestamp="2025-10-04 05:43:44 +0000 UTC" firstStartedPulling="2025-10-04 05:43:47.667249959 +0000 UTC m=+4178.995808773" lastFinishedPulling="2025-10-04 05:43:50.18025264 +0000 UTC m=+4181.508811454" observedRunningTime="2025-10-04 05:43:50.759368836 +0000 UTC m=+4182.087927670" watchObservedRunningTime="2025-10-04 05:43:50.760623662 +0000 UTC m=+4182.089182496" Oct 04 05:43:55 crc kubenswrapper[4575]: I1004 05:43:55.141377 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:55 crc kubenswrapper[4575]: I1004 05:43:55.144144 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:55 crc kubenswrapper[4575]: I1004 05:43:55.210348 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:55 crc kubenswrapper[4575]: I1004 05:43:55.833097 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:55 crc kubenswrapper[4575]: I1004 05:43:55.896096 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj28v"] Oct 04 05:43:57 crc kubenswrapper[4575]: I1004 05:43:57.776668 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nj28v" podUID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerName="registry-server" containerID="cri-o://1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf" gracePeriod=2 Oct 04 05:43:57 crc kubenswrapper[4575]: I1004 05:43:57.870535 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lgrd7"] Oct 04 05:43:57 crc kubenswrapper[4575]: I1004 05:43:57.873763 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:57 crc kubenswrapper[4575]: I1004 05:43:57.917369 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgrd7"] Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.067015 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-utilities\") pod \"redhat-operators-lgrd7\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.067178 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-catalog-content\") pod \"redhat-operators-lgrd7\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.067389 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cftph\" (UniqueName: \"kubernetes.io/projected/59b5c284-1dfc-466a-9ffa-37ed439d560d-kube-api-access-cftph\") pod \"redhat-operators-lgrd7\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.169711 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cftph\" (UniqueName: \"kubernetes.io/projected/59b5c284-1dfc-466a-9ffa-37ed439d560d-kube-api-access-cftph\") pod \"redhat-operators-lgrd7\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.169813 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-utilities\") pod \"redhat-operators-lgrd7\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.169913 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-catalog-content\") pod \"redhat-operators-lgrd7\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.170547 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-catalog-content\") pod \"redhat-operators-lgrd7\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.171219 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-utilities\") pod \"redhat-operators-lgrd7\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.206544 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cftph\" (UniqueName: \"kubernetes.io/projected/59b5c284-1dfc-466a-9ffa-37ed439d560d-kube-api-access-cftph\") pod \"redhat-operators-lgrd7\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.345942 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.491938 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.584447 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-catalog-content\") pod \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.585199 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-utilities\") pod \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.585291 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rlg4\" (UniqueName: \"kubernetes.io/projected/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-kube-api-access-5rlg4\") pod \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\" (UID: \"7f1c3aac-6b1f-48e4-846b-192e0977fb7a\") " Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.587451 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-utilities" (OuterVolumeSpecName: "utilities") pod "7f1c3aac-6b1f-48e4-846b-192e0977fb7a" (UID: "7f1c3aac-6b1f-48e4-846b-192e0977fb7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.592665 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-kube-api-access-5rlg4" (OuterVolumeSpecName: "kube-api-access-5rlg4") pod "7f1c3aac-6b1f-48e4-846b-192e0977fb7a" (UID: "7f1c3aac-6b1f-48e4-846b-192e0977fb7a"). InnerVolumeSpecName "kube-api-access-5rlg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.622104 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f1c3aac-6b1f-48e4-846b-192e0977fb7a" (UID: "7f1c3aac-6b1f-48e4-846b-192e0977fb7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.687835 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.687876 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.687891 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rlg4\" (UniqueName: \"kubernetes.io/projected/7f1c3aac-6b1f-48e4-846b-192e0977fb7a-kube-api-access-5rlg4\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.789843 4575 generic.go:334] "Generic (PLEG): container finished" podID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerID="1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf" exitCode=0 Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.789907 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj28v" event={"ID":"7f1c3aac-6b1f-48e4-846b-192e0977fb7a","Type":"ContainerDied","Data":"1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf"} Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.789959 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nj28v" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.789991 4575 scope.go:117] "RemoveContainer" containerID="1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.789967 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj28v" event={"ID":"7f1c3aac-6b1f-48e4-846b-192e0977fb7a","Type":"ContainerDied","Data":"ac6d7132c364bc0761d549247a0412e1e402012a64e256ed21456e8fb860712e"} Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.829088 4575 scope.go:117] "RemoveContainer" containerID="4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.834857 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj28v"] Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.844005 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj28v"] Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.860279 4575 scope.go:117] "RemoveContainer" containerID="7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.884188 4575 scope.go:117] "RemoveContainer" containerID="1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf" Oct 04 05:43:58 crc kubenswrapper[4575]: E1004 05:43:58.884724 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf\": container with ID starting with 1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf not found: ID does not exist" containerID="1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.884863 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf"} err="failed to get container status \"1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf\": rpc error: code = NotFound desc = could not find container \"1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf\": container with ID starting with 1232927433f5e54b3b12f6effd17a723950e67183abc01d2cc630e4930eedeaf not found: ID does not exist" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.884971 4575 scope.go:117] "RemoveContainer" containerID="4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488" Oct 04 05:43:58 crc kubenswrapper[4575]: E1004 05:43:58.885466 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488\": container with ID starting with 4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488 not found: ID does not exist" containerID="4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.885531 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488"} err="failed to get container status \"4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488\": rpc error: code = NotFound desc = could not find container \"4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488\": container with ID starting with 4d44443477729b716b28ac988ce8abf3a390ebf8c0567e90e4470cdfee252488 not found: ID does not exist" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.885563 4575 scope.go:117] "RemoveContainer" containerID="7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d" Oct 04 05:43:58 crc kubenswrapper[4575]: E1004 05:43:58.886042 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d\": container with ID starting with 7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d not found: ID does not exist" containerID="7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.886075 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d"} err="failed to get container status \"7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d\": rpc error: code = NotFound desc = could not find container \"7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d\": container with ID starting with 7c8a869aa67d0a547d85be04f21f4fb7797a0a9b663ce16d12e7f30dd65a073d not found: ID does not exist" Oct 04 05:43:58 crc kubenswrapper[4575]: I1004 05:43:58.922458 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgrd7"] Oct 04 05:43:59 crc kubenswrapper[4575]: I1004 05:43:59.338321 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" path="/var/lib/kubelet/pods/7f1c3aac-6b1f-48e4-846b-192e0977fb7a/volumes" Oct 04 05:43:59 crc kubenswrapper[4575]: I1004 05:43:59.800152 4575 generic.go:334] "Generic (PLEG): container finished" podID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerID="dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8" exitCode=0 Oct 04 05:43:59 crc kubenswrapper[4575]: I1004 05:43:59.800200 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrd7" event={"ID":"59b5c284-1dfc-466a-9ffa-37ed439d560d","Type":"ContainerDied","Data":"dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8"} Oct 04 05:43:59 crc kubenswrapper[4575]: I1004 05:43:59.800226 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrd7" event={"ID":"59b5c284-1dfc-466a-9ffa-37ed439d560d","Type":"ContainerStarted","Data":"56f5ea23be5577e7374bb827e21a8786552c87e8b781a3ed6216068841e26094"} Oct 04 05:44:01 crc kubenswrapper[4575]: I1004 05:44:01.843639 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrd7" event={"ID":"59b5c284-1dfc-466a-9ffa-37ed439d560d","Type":"ContainerStarted","Data":"986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999"} Oct 04 05:44:04 crc kubenswrapper[4575]: I1004 05:44:04.883814 4575 generic.go:334] "Generic (PLEG): container finished" podID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerID="986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999" exitCode=0 Oct 04 05:44:04 crc kubenswrapper[4575]: I1004 05:44:04.883856 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrd7" event={"ID":"59b5c284-1dfc-466a-9ffa-37ed439d560d","Type":"ContainerDied","Data":"986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999"} Oct 04 05:44:05 crc kubenswrapper[4575]: I1004 05:44:05.922913 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrd7" event={"ID":"59b5c284-1dfc-466a-9ffa-37ed439d560d","Type":"ContainerStarted","Data":"d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb"} Oct 04 05:44:05 crc kubenswrapper[4575]: I1004 05:44:05.975220 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lgrd7" podStartSLOduration=3.5247220280000002 podStartE2EDuration="8.975193257s" podCreationTimestamp="2025-10-04 05:43:57 +0000 UTC" firstStartedPulling="2025-10-04 05:43:59.803706242 +0000 UTC m=+4191.132265056" lastFinishedPulling="2025-10-04 05:44:05.254177471 +0000 UTC m=+4196.582736285" observedRunningTime="2025-10-04 05:44:05.950641428 +0000 UTC m=+4197.279200242" watchObservedRunningTime="2025-10-04 05:44:05.975193257 +0000 UTC m=+4197.303752071" Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.347304 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.348723 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.446926 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.447000 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.447052 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.447900 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8dce56ed3e369405293833fa9a23c605f7ddc95e08b651fd4ef5764d7d1df662"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.447964 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://8dce56ed3e369405293833fa9a23c605f7ddc95e08b651fd4ef5764d7d1df662" gracePeriod=600 Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.956707 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="8dce56ed3e369405293833fa9a23c605f7ddc95e08b651fd4ef5764d7d1df662" exitCode=0 Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.958697 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"8dce56ed3e369405293833fa9a23c605f7ddc95e08b651fd4ef5764d7d1df662"} Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.958751 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e"} Oct 04 05:44:08 crc kubenswrapper[4575]: I1004 05:44:08.958777 4575 scope.go:117] "RemoveContainer" containerID="1a1402785ce5c77c9b7a8e468335d4bc155ea35724d20aec1adecff2c0cb0087" Oct 04 05:44:09 crc kubenswrapper[4575]: I1004 05:44:09.411199 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lgrd7" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerName="registry-server" probeResult="failure" output=< Oct 04 05:44:09 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:44:09 crc kubenswrapper[4575]: > Oct 04 05:44:18 crc kubenswrapper[4575]: I1004 05:44:18.404792 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:44:18 crc kubenswrapper[4575]: I1004 05:44:18.483906 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:44:18 crc kubenswrapper[4575]: I1004 05:44:18.651899 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgrd7"] Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.084368 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lgrd7" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerName="registry-server" containerID="cri-o://d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb" gracePeriod=2 Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.678346 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.843111 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-catalog-content\") pod \"59b5c284-1dfc-466a-9ffa-37ed439d560d\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.843459 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cftph\" (UniqueName: \"kubernetes.io/projected/59b5c284-1dfc-466a-9ffa-37ed439d560d-kube-api-access-cftph\") pod \"59b5c284-1dfc-466a-9ffa-37ed439d560d\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.843734 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-utilities\") pod \"59b5c284-1dfc-466a-9ffa-37ed439d560d\" (UID: \"59b5c284-1dfc-466a-9ffa-37ed439d560d\") " Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.844386 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-utilities" (OuterVolumeSpecName: "utilities") pod "59b5c284-1dfc-466a-9ffa-37ed439d560d" (UID: "59b5c284-1dfc-466a-9ffa-37ed439d560d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.844689 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.851006 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b5c284-1dfc-466a-9ffa-37ed439d560d-kube-api-access-cftph" (OuterVolumeSpecName: "kube-api-access-cftph") pod "59b5c284-1dfc-466a-9ffa-37ed439d560d" (UID: "59b5c284-1dfc-466a-9ffa-37ed439d560d"). InnerVolumeSpecName "kube-api-access-cftph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.947027 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cftph\" (UniqueName: \"kubernetes.io/projected/59b5c284-1dfc-466a-9ffa-37ed439d560d-kube-api-access-cftph\") on node \"crc\" DevicePath \"\"" Oct 04 05:44:20 crc kubenswrapper[4575]: I1004 05:44:20.951891 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59b5c284-1dfc-466a-9ffa-37ed439d560d" (UID: "59b5c284-1dfc-466a-9ffa-37ed439d560d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.049324 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b5c284-1dfc-466a-9ffa-37ed439d560d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.096951 4575 generic.go:334] "Generic (PLEG): container finished" podID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerID="d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb" exitCode=0 Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.096991 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrd7" event={"ID":"59b5c284-1dfc-466a-9ffa-37ed439d560d","Type":"ContainerDied","Data":"d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb"} Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.097058 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrd7" event={"ID":"59b5c284-1dfc-466a-9ffa-37ed439d560d","Type":"ContainerDied","Data":"56f5ea23be5577e7374bb827e21a8786552c87e8b781a3ed6216068841e26094"} Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.097084 4575 scope.go:117] "RemoveContainer" containerID="d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.097089 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgrd7" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.131426 4575 scope.go:117] "RemoveContainer" containerID="986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.156710 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgrd7"] Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.167882 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lgrd7"] Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.188446 4575 scope.go:117] "RemoveContainer" containerID="dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.242868 4575 scope.go:117] "RemoveContainer" containerID="d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb" Oct 04 05:44:21 crc kubenswrapper[4575]: E1004 05:44:21.243596 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb\": container with ID starting with d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb not found: ID does not exist" containerID="d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.243651 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb"} err="failed to get container status \"d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb\": rpc error: code = NotFound desc = could not find container \"d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb\": container with ID starting with d2604c57cae4edac5be48c13c64239035a18d29db5c49c836b5efd1bd6652dcb not found: ID does not exist" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.243683 4575 scope.go:117] "RemoveContainer" containerID="986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999" Oct 04 05:44:21 crc kubenswrapper[4575]: E1004 05:44:21.244284 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999\": container with ID starting with 986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999 not found: ID does not exist" containerID="986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.244310 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999"} err="failed to get container status \"986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999\": rpc error: code = NotFound desc = could not find container \"986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999\": container with ID starting with 986b24791167ce47fd578cc49c11330028909930f2c7590fe6ffc51de468e999 not found: ID does not exist" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.244327 4575 scope.go:117] "RemoveContainer" containerID="dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8" Oct 04 05:44:21 crc kubenswrapper[4575]: E1004 05:44:21.244942 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8\": container with ID starting with dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8 not found: ID does not exist" containerID="dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.244982 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8"} err="failed to get container status \"dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8\": rpc error: code = NotFound desc = could not find container \"dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8\": container with ID starting with dfdb0bda1f253e3e38724f6256c9abe8a9ab3a8f9213dbc7af1a9adcce2bcfe8 not found: ID does not exist" Oct 04 05:44:21 crc kubenswrapper[4575]: I1004 05:44:21.324745 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" path="/var/lib/kubelet/pods/59b5c284-1dfc-466a-9ffa-37ed439d560d/volumes" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.155709 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26"] Oct 04 05:45:00 crc kubenswrapper[4575]: E1004 05:45:00.156781 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerName="extract-content" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.156801 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerName="extract-content" Oct 04 05:45:00 crc kubenswrapper[4575]: E1004 05:45:00.156821 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerName="extract-utilities" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.156829 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerName="extract-utilities" Oct 04 05:45:00 crc kubenswrapper[4575]: E1004 05:45:00.156848 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerName="extract-utilities" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.156855 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerName="extract-utilities" Oct 04 05:45:00 crc kubenswrapper[4575]: E1004 05:45:00.156916 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerName="extract-content" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.156925 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerName="extract-content" Oct 04 05:45:00 crc kubenswrapper[4575]: E1004 05:45:00.156941 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.156948 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4575]: E1004 05:45:00.156973 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.156981 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.157299 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f1c3aac-6b1f-48e4-846b-192e0977fb7a" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.157320 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b5c284-1dfc-466a-9ffa-37ed439d560d" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.158184 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.160045 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.162578 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.175518 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26"] Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.265894 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-secret-volume\") pod \"collect-profiles-29325945-52c26\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.266305 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-config-volume\") pod \"collect-profiles-29325945-52c26\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.266361 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6cw4\" (UniqueName: \"kubernetes.io/projected/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-kube-api-access-v6cw4\") pod \"collect-profiles-29325945-52c26\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.368294 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-secret-volume\") pod \"collect-profiles-29325945-52c26\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.368362 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-config-volume\") pod \"collect-profiles-29325945-52c26\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.368411 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6cw4\" (UniqueName: \"kubernetes.io/projected/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-kube-api-access-v6cw4\") pod \"collect-profiles-29325945-52c26\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.371178 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-config-volume\") pod \"collect-profiles-29325945-52c26\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.377434 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-secret-volume\") pod \"collect-profiles-29325945-52c26\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.390798 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6cw4\" (UniqueName: \"kubernetes.io/projected/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-kube-api-access-v6cw4\") pod \"collect-profiles-29325945-52c26\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.487546 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:00 crc kubenswrapper[4575]: I1004 05:45:00.998664 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26"] Oct 04 05:45:01 crc kubenswrapper[4575]: I1004 05:45:01.511975 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" event={"ID":"cace4b18-ad5b-4a58-89ad-36bd6ca87dda","Type":"ContainerStarted","Data":"00c427b5f00013af31115612933605c126c4e7a0b51ffac7b53b97a52c606da2"} Oct 04 05:45:01 crc kubenswrapper[4575]: I1004 05:45:01.512737 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" event={"ID":"cace4b18-ad5b-4a58-89ad-36bd6ca87dda","Type":"ContainerStarted","Data":"6600e6330e8b79ce85d244d4aa49302b863b797da897786d65d5f86e8783a313"} Oct 04 05:45:01 crc kubenswrapper[4575]: I1004 05:45:01.535070 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" podStartSLOduration=1.535049154 podStartE2EDuration="1.535049154s" podCreationTimestamp="2025-10-04 05:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:45:01.531006188 +0000 UTC m=+4252.859565012" watchObservedRunningTime="2025-10-04 05:45:01.535049154 +0000 UTC m=+4252.863607978" Oct 04 05:45:02 crc kubenswrapper[4575]: I1004 05:45:02.527011 4575 generic.go:334] "Generic (PLEG): container finished" podID="cace4b18-ad5b-4a58-89ad-36bd6ca87dda" containerID="00c427b5f00013af31115612933605c126c4e7a0b51ffac7b53b97a52c606da2" exitCode=0 Oct 04 05:45:02 crc kubenswrapper[4575]: I1004 05:45:02.527884 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" event={"ID":"cace4b18-ad5b-4a58-89ad-36bd6ca87dda","Type":"ContainerDied","Data":"00c427b5f00013af31115612933605c126c4e7a0b51ffac7b53b97a52c606da2"} Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.008483 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.066053 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-secret-volume\") pod \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.066270 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6cw4\" (UniqueName: \"kubernetes.io/projected/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-kube-api-access-v6cw4\") pod \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.066323 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-config-volume\") pod \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\" (UID: \"cace4b18-ad5b-4a58-89ad-36bd6ca87dda\") " Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.067833 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-config-volume" (OuterVolumeSpecName: "config-volume") pod "cace4b18-ad5b-4a58-89ad-36bd6ca87dda" (UID: "cace4b18-ad5b-4a58-89ad-36bd6ca87dda"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.075679 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-kube-api-access-v6cw4" (OuterVolumeSpecName: "kube-api-access-v6cw4") pod "cace4b18-ad5b-4a58-89ad-36bd6ca87dda" (UID: "cace4b18-ad5b-4a58-89ad-36bd6ca87dda"). InnerVolumeSpecName "kube-api-access-v6cw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.087321 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cace4b18-ad5b-4a58-89ad-36bd6ca87dda" (UID: "cace4b18-ad5b-4a58-89ad-36bd6ca87dda"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.168558 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6cw4\" (UniqueName: \"kubernetes.io/projected/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-kube-api-access-v6cw4\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.168618 4575 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.168627 4575 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cace4b18-ad5b-4a58-89ad-36bd6ca87dda-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.562200 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" event={"ID":"cace4b18-ad5b-4a58-89ad-36bd6ca87dda","Type":"ContainerDied","Data":"6600e6330e8b79ce85d244d4aa49302b863b797da897786d65d5f86e8783a313"} Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.562259 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6600e6330e8b79ce85d244d4aa49302b863b797da897786d65d5f86e8783a313" Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.562367 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-52c26" Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.625903 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8"] Oct 04 05:45:04 crc kubenswrapper[4575]: I1004 05:45:04.633843 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-fkqd8"] Oct 04 05:45:05 crc kubenswrapper[4575]: I1004 05:45:05.320614 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b88d73-0f40-457f-8ece-5cdb2b87a0ec" path="/var/lib/kubelet/pods/b9b88d73-0f40-457f-8ece-5cdb2b87a0ec/volumes" Oct 04 05:45:31 crc kubenswrapper[4575]: I1004 05:45:31.739336 4575 scope.go:117] "RemoveContainer" containerID="8dc871af7f20ebe706d30ede36fcfbc471d80d3cd68dbdc0ba13457f4c0b631f" Oct 04 05:46:08 crc kubenswrapper[4575]: I1004 05:46:08.446347 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:46:08 crc kubenswrapper[4575]: I1004 05:46:08.446973 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:46:38 crc kubenswrapper[4575]: I1004 05:46:38.446074 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:46:38 crc kubenswrapper[4575]: I1004 05:46:38.446804 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:47:08 crc kubenswrapper[4575]: I1004 05:47:08.446694 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:47:08 crc kubenswrapper[4575]: I1004 05:47:08.447402 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:47:08 crc kubenswrapper[4575]: I1004 05:47:08.447502 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:47:08 crc kubenswrapper[4575]: I1004 05:47:08.448997 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:47:08 crc kubenswrapper[4575]: I1004 05:47:08.449082 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" gracePeriod=600 Oct 04 05:47:08 crc kubenswrapper[4575]: E1004 05:47:08.575957 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:47:08 crc kubenswrapper[4575]: I1004 05:47:08.661962 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" exitCode=0 Oct 04 05:47:08 crc kubenswrapper[4575]: I1004 05:47:08.662028 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e"} Oct 04 05:47:08 crc kubenswrapper[4575]: I1004 05:47:08.662069 4575 scope.go:117] "RemoveContainer" containerID="8dce56ed3e369405293833fa9a23c605f7ddc95e08b651fd4ef5764d7d1df662" Oct 04 05:47:08 crc kubenswrapper[4575]: I1004 05:47:08.662899 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:47:08 crc kubenswrapper[4575]: E1004 05:47:08.663147 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:47:21 crc kubenswrapper[4575]: I1004 05:47:21.311601 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:47:21 crc kubenswrapper[4575]: E1004 05:47:21.312368 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.634681 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lsx26"] Oct 04 05:47:23 crc kubenswrapper[4575]: E1004 05:47:23.635649 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cace4b18-ad5b-4a58-89ad-36bd6ca87dda" containerName="collect-profiles" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.635663 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="cace4b18-ad5b-4a58-89ad-36bd6ca87dda" containerName="collect-profiles" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.635880 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="cace4b18-ad5b-4a58-89ad-36bd6ca87dda" containerName="collect-profiles" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.637248 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.653813 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-utilities\") pod \"community-operators-lsx26\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.653926 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-catalog-content\") pod \"community-operators-lsx26\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.654280 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lzcj\" (UniqueName: \"kubernetes.io/projected/e124bf9c-d964-412d-8ca9-5474617cf841-kube-api-access-7lzcj\") pod \"community-operators-lsx26\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.655079 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lsx26"] Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.756009 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-utilities\") pod \"community-operators-lsx26\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.756125 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-catalog-content\") pod \"community-operators-lsx26\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.756224 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lzcj\" (UniqueName: \"kubernetes.io/projected/e124bf9c-d964-412d-8ca9-5474617cf841-kube-api-access-7lzcj\") pod \"community-operators-lsx26\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.756731 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-utilities\") pod \"community-operators-lsx26\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.756785 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-catalog-content\") pod \"community-operators-lsx26\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.779497 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lzcj\" (UniqueName: \"kubernetes.io/projected/e124bf9c-d964-412d-8ca9-5474617cf841-kube-api-access-7lzcj\") pod \"community-operators-lsx26\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:23 crc kubenswrapper[4575]: I1004 05:47:23.963826 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:24 crc kubenswrapper[4575]: I1004 05:47:24.551402 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lsx26"] Oct 04 05:47:24 crc kubenswrapper[4575]: I1004 05:47:24.861118 4575 generic.go:334] "Generic (PLEG): container finished" podID="e124bf9c-d964-412d-8ca9-5474617cf841" containerID="2e29ca52aca228b82b97c435a9eb119a19ce90e65be22db3660cab0d4ffe1457" exitCode=0 Oct 04 05:47:24 crc kubenswrapper[4575]: I1004 05:47:24.861183 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lsx26" event={"ID":"e124bf9c-d964-412d-8ca9-5474617cf841","Type":"ContainerDied","Data":"2e29ca52aca228b82b97c435a9eb119a19ce90e65be22db3660cab0d4ffe1457"} Oct 04 05:47:24 crc kubenswrapper[4575]: I1004 05:47:24.861521 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lsx26" event={"ID":"e124bf9c-d964-412d-8ca9-5474617cf841","Type":"ContainerStarted","Data":"11ca8b34a2d7d80dd74f14bde4d6af809739e242203ba1565b3483708bdf5d62"} Oct 04 05:47:25 crc kubenswrapper[4575]: I1004 05:47:25.874376 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lsx26" event={"ID":"e124bf9c-d964-412d-8ca9-5474617cf841","Type":"ContainerStarted","Data":"d6b00ed755b25e0df9affd27a469dddc9ee47cd6397200e2ef8fa1bd6deab4e7"} Oct 04 05:47:26 crc kubenswrapper[4575]: I1004 05:47:26.884893 4575 generic.go:334] "Generic (PLEG): container finished" podID="e124bf9c-d964-412d-8ca9-5474617cf841" containerID="d6b00ed755b25e0df9affd27a469dddc9ee47cd6397200e2ef8fa1bd6deab4e7" exitCode=0 Oct 04 05:47:26 crc kubenswrapper[4575]: I1004 05:47:26.885253 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lsx26" event={"ID":"e124bf9c-d964-412d-8ca9-5474617cf841","Type":"ContainerDied","Data":"d6b00ed755b25e0df9affd27a469dddc9ee47cd6397200e2ef8fa1bd6deab4e7"} Oct 04 05:47:27 crc kubenswrapper[4575]: I1004 05:47:27.899748 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lsx26" event={"ID":"e124bf9c-d964-412d-8ca9-5474617cf841","Type":"ContainerStarted","Data":"369a49c4a86d4459f042c87edfcbace1eef1c7efea38bbcdaa2359df79e47c84"} Oct 04 05:47:27 crc kubenswrapper[4575]: I1004 05:47:27.925842 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lsx26" podStartSLOduration=2.379612294 podStartE2EDuration="4.925822276s" podCreationTimestamp="2025-10-04 05:47:23 +0000 UTC" firstStartedPulling="2025-10-04 05:47:24.863175161 +0000 UTC m=+4396.191733975" lastFinishedPulling="2025-10-04 05:47:27.409385143 +0000 UTC m=+4398.737943957" observedRunningTime="2025-10-04 05:47:27.922386237 +0000 UTC m=+4399.250945071" watchObservedRunningTime="2025-10-04 05:47:27.925822276 +0000 UTC m=+4399.254381100" Oct 04 05:47:33 crc kubenswrapper[4575]: I1004 05:47:33.964304 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:33 crc kubenswrapper[4575]: I1004 05:47:33.966725 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:34 crc kubenswrapper[4575]: I1004 05:47:34.011867 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:34 crc kubenswrapper[4575]: I1004 05:47:34.309909 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:47:34 crc kubenswrapper[4575]: E1004 05:47:34.311088 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:47:35 crc kubenswrapper[4575]: I1004 05:47:35.010176 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:35 crc kubenswrapper[4575]: I1004 05:47:35.060629 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lsx26"] Oct 04 05:47:36 crc kubenswrapper[4575]: I1004 05:47:36.975336 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lsx26" podUID="e124bf9c-d964-412d-8ca9-5474617cf841" containerName="registry-server" containerID="cri-o://369a49c4a86d4459f042c87edfcbace1eef1c7efea38bbcdaa2359df79e47c84" gracePeriod=2 Oct 04 05:47:37 crc kubenswrapper[4575]: I1004 05:47:37.993014 4575 generic.go:334] "Generic (PLEG): container finished" podID="e124bf9c-d964-412d-8ca9-5474617cf841" containerID="369a49c4a86d4459f042c87edfcbace1eef1c7efea38bbcdaa2359df79e47c84" exitCode=0 Oct 04 05:47:37 crc kubenswrapper[4575]: I1004 05:47:37.993224 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lsx26" event={"ID":"e124bf9c-d964-412d-8ca9-5474617cf841","Type":"ContainerDied","Data":"369a49c4a86d4459f042c87edfcbace1eef1c7efea38bbcdaa2359df79e47c84"} Oct 04 05:47:37 crc kubenswrapper[4575]: I1004 05:47:37.993295 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lsx26" event={"ID":"e124bf9c-d964-412d-8ca9-5474617cf841","Type":"ContainerDied","Data":"11ca8b34a2d7d80dd74f14bde4d6af809739e242203ba1565b3483708bdf5d62"} Oct 04 05:47:37 crc kubenswrapper[4575]: I1004 05:47:37.993307 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11ca8b34a2d7d80dd74f14bde4d6af809739e242203ba1565b3483708bdf5d62" Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.070760 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.148900 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lzcj\" (UniqueName: \"kubernetes.io/projected/e124bf9c-d964-412d-8ca9-5474617cf841-kube-api-access-7lzcj\") pod \"e124bf9c-d964-412d-8ca9-5474617cf841\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.149001 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-utilities\") pod \"e124bf9c-d964-412d-8ca9-5474617cf841\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.149182 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-catalog-content\") pod \"e124bf9c-d964-412d-8ca9-5474617cf841\" (UID: \"e124bf9c-d964-412d-8ca9-5474617cf841\") " Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.150750 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-utilities" (OuterVolumeSpecName: "utilities") pod "e124bf9c-d964-412d-8ca9-5474617cf841" (UID: "e124bf9c-d964-412d-8ca9-5474617cf841"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.168134 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e124bf9c-d964-412d-8ca9-5474617cf841-kube-api-access-7lzcj" (OuterVolumeSpecName: "kube-api-access-7lzcj") pod "e124bf9c-d964-412d-8ca9-5474617cf841" (UID: "e124bf9c-d964-412d-8ca9-5474617cf841"). InnerVolumeSpecName "kube-api-access-7lzcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.212611 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e124bf9c-d964-412d-8ca9-5474617cf841" (UID: "e124bf9c-d964-412d-8ca9-5474617cf841"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.250273 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.250323 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lzcj\" (UniqueName: \"kubernetes.io/projected/e124bf9c-d964-412d-8ca9-5474617cf841-kube-api-access-7lzcj\") on node \"crc\" DevicePath \"\"" Oct 04 05:47:38 crc kubenswrapper[4575]: I1004 05:47:38.250343 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e124bf9c-d964-412d-8ca9-5474617cf841-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:47:39 crc kubenswrapper[4575]: I1004 05:47:39.001303 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lsx26" Oct 04 05:47:39 crc kubenswrapper[4575]: I1004 05:47:39.045032 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lsx26"] Oct 04 05:47:39 crc kubenswrapper[4575]: I1004 05:47:39.053840 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lsx26"] Oct 04 05:47:39 crc kubenswrapper[4575]: I1004 05:47:39.320900 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e124bf9c-d964-412d-8ca9-5474617cf841" path="/var/lib/kubelet/pods/e124bf9c-d964-412d-8ca9-5474617cf841/volumes" Oct 04 05:47:48 crc kubenswrapper[4575]: I1004 05:47:48.310167 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:47:48 crc kubenswrapper[4575]: E1004 05:47:48.311100 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:48:01 crc kubenswrapper[4575]: I1004 05:48:01.309852 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:48:01 crc kubenswrapper[4575]: E1004 05:48:01.310604 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:48:16 crc kubenswrapper[4575]: I1004 05:48:16.309701 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:48:16 crc kubenswrapper[4575]: E1004 05:48:16.310645 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:48:28 crc kubenswrapper[4575]: I1004 05:48:28.310279 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:48:28 crc kubenswrapper[4575]: E1004 05:48:28.310988 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:48:43 crc kubenswrapper[4575]: I1004 05:48:43.311132 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:48:43 crc kubenswrapper[4575]: E1004 05:48:43.312421 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:48:56 crc kubenswrapper[4575]: I1004 05:48:56.310908 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:48:56 crc kubenswrapper[4575]: E1004 05:48:56.311648 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:49:10 crc kubenswrapper[4575]: I1004 05:49:10.309960 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:49:10 crc kubenswrapper[4575]: E1004 05:49:10.310785 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:49:23 crc kubenswrapper[4575]: I1004 05:49:23.310400 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:49:23 crc kubenswrapper[4575]: E1004 05:49:23.311209 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:49:37 crc kubenswrapper[4575]: I1004 05:49:37.310281 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:49:37 crc kubenswrapper[4575]: E1004 05:49:37.310992 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:49:51 crc kubenswrapper[4575]: I1004 05:49:51.310314 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:49:51 crc kubenswrapper[4575]: E1004 05:49:51.311036 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:50:05 crc kubenswrapper[4575]: I1004 05:50:05.310389 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:50:05 crc kubenswrapper[4575]: E1004 05:50:05.311185 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:50:18 crc kubenswrapper[4575]: I1004 05:50:18.311306 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:50:18 crc kubenswrapper[4575]: E1004 05:50:18.312259 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:50:32 crc kubenswrapper[4575]: I1004 05:50:32.310486 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:50:32 crc kubenswrapper[4575]: E1004 05:50:32.311344 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:50:47 crc kubenswrapper[4575]: I1004 05:50:47.310711 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:50:47 crc kubenswrapper[4575]: E1004 05:50:47.311797 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:50:58 crc kubenswrapper[4575]: I1004 05:50:58.310965 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:50:58 crc kubenswrapper[4575]: E1004 05:50:58.312280 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:51:13 crc kubenswrapper[4575]: I1004 05:51:13.311103 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:51:13 crc kubenswrapper[4575]: E1004 05:51:13.312292 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:51:28 crc kubenswrapper[4575]: I1004 05:51:28.310786 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:51:28 crc kubenswrapper[4575]: E1004 05:51:28.311623 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:51:40 crc kubenswrapper[4575]: I1004 05:51:40.310217 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:51:40 crc kubenswrapper[4575]: E1004 05:51:40.310929 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:51:53 crc kubenswrapper[4575]: I1004 05:51:53.310695 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:51:53 crc kubenswrapper[4575]: E1004 05:51:53.311466 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:52:07 crc kubenswrapper[4575]: I1004 05:52:07.314403 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:52:07 crc kubenswrapper[4575]: E1004 05:52:07.315149 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:52:20 crc kubenswrapper[4575]: I1004 05:52:20.309807 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:52:20 crc kubenswrapper[4575]: I1004 05:52:20.555364 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"c52f1c24086c003b61d72bbb37cb4c73370308383293e59bd9c3a69ae525e0d4"} Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.098677 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f5zg5"] Oct 04 05:53:19 crc kubenswrapper[4575]: E1004 05:53:19.099930 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e124bf9c-d964-412d-8ca9-5474617cf841" containerName="registry-server" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.099950 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="e124bf9c-d964-412d-8ca9-5474617cf841" containerName="registry-server" Oct 04 05:53:19 crc kubenswrapper[4575]: E1004 05:53:19.099968 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e124bf9c-d964-412d-8ca9-5474617cf841" containerName="extract-content" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.099975 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="e124bf9c-d964-412d-8ca9-5474617cf841" containerName="extract-content" Oct 04 05:53:19 crc kubenswrapper[4575]: E1004 05:53:19.100024 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e124bf9c-d964-412d-8ca9-5474617cf841" containerName="extract-utilities" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.100033 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="e124bf9c-d964-412d-8ca9-5474617cf841" containerName="extract-utilities" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.100289 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="e124bf9c-d964-412d-8ca9-5474617cf841" containerName="registry-server" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.101997 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.114374 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f5zg5"] Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.237529 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjhcw\" (UniqueName: \"kubernetes.io/projected/1dedc78d-f8c9-468e-bec8-b0b16d93161c-kube-api-access-jjhcw\") pod \"certified-operators-f5zg5\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.237660 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-utilities\") pod \"certified-operators-f5zg5\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.237711 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-catalog-content\") pod \"certified-operators-f5zg5\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.339595 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjhcw\" (UniqueName: \"kubernetes.io/projected/1dedc78d-f8c9-468e-bec8-b0b16d93161c-kube-api-access-jjhcw\") pod \"certified-operators-f5zg5\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.339714 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-utilities\") pod \"certified-operators-f5zg5\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.339773 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-catalog-content\") pod \"certified-operators-f5zg5\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.340483 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-catalog-content\") pod \"certified-operators-f5zg5\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.341158 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-utilities\") pod \"certified-operators-f5zg5\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.376807 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjhcw\" (UniqueName: \"kubernetes.io/projected/1dedc78d-f8c9-468e-bec8-b0b16d93161c-kube-api-access-jjhcw\") pod \"certified-operators-f5zg5\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:19 crc kubenswrapper[4575]: I1004 05:53:19.427672 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:20 crc kubenswrapper[4575]: I1004 05:53:20.221818 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f5zg5"] Oct 04 05:53:20 crc kubenswrapper[4575]: W1004 05:53:20.225839 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1dedc78d_f8c9_468e_bec8_b0b16d93161c.slice/crio-78c2de8aceb3e1351af4f9b01d99353efebd82c5a0a7fa5bdff8d851967f488f WatchSource:0}: Error finding container 78c2de8aceb3e1351af4f9b01d99353efebd82c5a0a7fa5bdff8d851967f488f: Status 404 returned error can't find the container with id 78c2de8aceb3e1351af4f9b01d99353efebd82c5a0a7fa5bdff8d851967f488f Oct 04 05:53:21 crc kubenswrapper[4575]: I1004 05:53:21.139713 4575 generic.go:334] "Generic (PLEG): container finished" podID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerID="6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1" exitCode=0 Oct 04 05:53:21 crc kubenswrapper[4575]: I1004 05:53:21.139817 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5zg5" event={"ID":"1dedc78d-f8c9-468e-bec8-b0b16d93161c","Type":"ContainerDied","Data":"6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1"} Oct 04 05:53:21 crc kubenswrapper[4575]: I1004 05:53:21.140110 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5zg5" event={"ID":"1dedc78d-f8c9-468e-bec8-b0b16d93161c","Type":"ContainerStarted","Data":"78c2de8aceb3e1351af4f9b01d99353efebd82c5a0a7fa5bdff8d851967f488f"} Oct 04 05:53:21 crc kubenswrapper[4575]: I1004 05:53:21.142642 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:53:23 crc kubenswrapper[4575]: I1004 05:53:23.158403 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5zg5" event={"ID":"1dedc78d-f8c9-468e-bec8-b0b16d93161c","Type":"ContainerStarted","Data":"2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5"} Oct 04 05:53:24 crc kubenswrapper[4575]: I1004 05:53:24.172059 4575 generic.go:334] "Generic (PLEG): container finished" podID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerID="2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5" exitCode=0 Oct 04 05:53:24 crc kubenswrapper[4575]: I1004 05:53:24.172109 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5zg5" event={"ID":"1dedc78d-f8c9-468e-bec8-b0b16d93161c","Type":"ContainerDied","Data":"2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5"} Oct 04 05:53:25 crc kubenswrapper[4575]: I1004 05:53:25.181741 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5zg5" event={"ID":"1dedc78d-f8c9-468e-bec8-b0b16d93161c","Type":"ContainerStarted","Data":"70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1"} Oct 04 05:53:25 crc kubenswrapper[4575]: I1004 05:53:25.203777 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f5zg5" podStartSLOduration=2.660402964 podStartE2EDuration="6.203758444s" podCreationTimestamp="2025-10-04 05:53:19 +0000 UTC" firstStartedPulling="2025-10-04 05:53:21.142387381 +0000 UTC m=+4752.470946195" lastFinishedPulling="2025-10-04 05:53:24.685742861 +0000 UTC m=+4756.014301675" observedRunningTime="2025-10-04 05:53:25.201621282 +0000 UTC m=+4756.530180116" watchObservedRunningTime="2025-10-04 05:53:25.203758444 +0000 UTC m=+4756.532317258" Oct 04 05:53:29 crc kubenswrapper[4575]: I1004 05:53:29.428895 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:29 crc kubenswrapper[4575]: I1004 05:53:29.429314 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:29 crc kubenswrapper[4575]: I1004 05:53:29.474795 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:30 crc kubenswrapper[4575]: I1004 05:53:30.279984 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:30 crc kubenswrapper[4575]: I1004 05:53:30.323697 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f5zg5"] Oct 04 05:53:31 crc kubenswrapper[4575]: I1004 05:53:31.982378 4575 scope.go:117] "RemoveContainer" containerID="2e29ca52aca228b82b97c435a9eb119a19ce90e65be22db3660cab0d4ffe1457" Oct 04 05:53:32 crc kubenswrapper[4575]: I1004 05:53:32.010422 4575 scope.go:117] "RemoveContainer" containerID="d6b00ed755b25e0df9affd27a469dddc9ee47cd6397200e2ef8fa1bd6deab4e7" Oct 04 05:53:32 crc kubenswrapper[4575]: I1004 05:53:32.061816 4575 scope.go:117] "RemoveContainer" containerID="369a49c4a86d4459f042c87edfcbace1eef1c7efea38bbcdaa2359df79e47c84" Oct 04 05:53:32 crc kubenswrapper[4575]: I1004 05:53:32.245064 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f5zg5" podUID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerName="registry-server" containerID="cri-o://70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1" gracePeriod=2 Oct 04 05:53:32 crc kubenswrapper[4575]: I1004 05:53:32.727682 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:32 crc kubenswrapper[4575]: I1004 05:53:32.912473 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-catalog-content\") pod \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " Oct 04 05:53:32 crc kubenswrapper[4575]: I1004 05:53:32.912946 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-utilities\") pod \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " Oct 04 05:53:32 crc kubenswrapper[4575]: I1004 05:53:32.912984 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjhcw\" (UniqueName: \"kubernetes.io/projected/1dedc78d-f8c9-468e-bec8-b0b16d93161c-kube-api-access-jjhcw\") pod \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\" (UID: \"1dedc78d-f8c9-468e-bec8-b0b16d93161c\") " Oct 04 05:53:32 crc kubenswrapper[4575]: I1004 05:53:32.913816 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-utilities" (OuterVolumeSpecName: "utilities") pod "1dedc78d-f8c9-468e-bec8-b0b16d93161c" (UID: "1dedc78d-f8c9-468e-bec8-b0b16d93161c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:53:32 crc kubenswrapper[4575]: I1004 05:53:32.918757 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dedc78d-f8c9-468e-bec8-b0b16d93161c-kube-api-access-jjhcw" (OuterVolumeSpecName: "kube-api-access-jjhcw") pod "1dedc78d-f8c9-468e-bec8-b0b16d93161c" (UID: "1dedc78d-f8c9-468e-bec8-b0b16d93161c"). InnerVolumeSpecName "kube-api-access-jjhcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.015120 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.015149 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjhcw\" (UniqueName: \"kubernetes.io/projected/1dedc78d-f8c9-468e-bec8-b0b16d93161c-kube-api-access-jjhcw\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.029012 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1dedc78d-f8c9-468e-bec8-b0b16d93161c" (UID: "1dedc78d-f8c9-468e-bec8-b0b16d93161c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.117547 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dedc78d-f8c9-468e-bec8-b0b16d93161c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.256552 4575 generic.go:334] "Generic (PLEG): container finished" podID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerID="70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1" exitCode=0 Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.256624 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5zg5" event={"ID":"1dedc78d-f8c9-468e-bec8-b0b16d93161c","Type":"ContainerDied","Data":"70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1"} Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.256677 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5zg5" event={"ID":"1dedc78d-f8c9-468e-bec8-b0b16d93161c","Type":"ContainerDied","Data":"78c2de8aceb3e1351af4f9b01d99353efebd82c5a0a7fa5bdff8d851967f488f"} Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.256698 4575 scope.go:117] "RemoveContainer" containerID="70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.257011 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f5zg5" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.278388 4575 scope.go:117] "RemoveContainer" containerID="2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.306794 4575 scope.go:117] "RemoveContainer" containerID="6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.332968 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f5zg5"] Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.333012 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f5zg5"] Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.360009 4575 scope.go:117] "RemoveContainer" containerID="70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1" Oct 04 05:53:33 crc kubenswrapper[4575]: E1004 05:53:33.360324 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1\": container with ID starting with 70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1 not found: ID does not exist" containerID="70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.360354 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1"} err="failed to get container status \"70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1\": rpc error: code = NotFound desc = could not find container \"70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1\": container with ID starting with 70c9bb21033d2ecf0eb5e2d2a2fccd6ea511aea3a9e5959c1d54e464561e9ff1 not found: ID does not exist" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.360373 4575 scope.go:117] "RemoveContainer" containerID="2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5" Oct 04 05:53:33 crc kubenswrapper[4575]: E1004 05:53:33.360574 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5\": container with ID starting with 2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5 not found: ID does not exist" containerID="2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.360632 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5"} err="failed to get container status \"2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5\": rpc error: code = NotFound desc = could not find container \"2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5\": container with ID starting with 2854d9f249d76b4fe4f0e4e47e5cb3464e1f988eb7c2ce25b94167c8dfd01ed5 not found: ID does not exist" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.360647 4575 scope.go:117] "RemoveContainer" containerID="6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1" Oct 04 05:53:33 crc kubenswrapper[4575]: E1004 05:53:33.361020 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1\": container with ID starting with 6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1 not found: ID does not exist" containerID="6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1" Oct 04 05:53:33 crc kubenswrapper[4575]: I1004 05:53:33.361043 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1"} err="failed to get container status \"6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1\": rpc error: code = NotFound desc = could not find container \"6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1\": container with ID starting with 6a3d6fbb7746565b494e0d70297c502f7c04af92212d365db5ac262a7c3c6bb1 not found: ID does not exist" Oct 04 05:53:35 crc kubenswrapper[4575]: I1004 05:53:35.320658 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" path="/var/lib/kubelet/pods/1dedc78d-f8c9-468e-bec8-b0b16d93161c/volumes" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.425554 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ljvqv"] Oct 04 05:54:02 crc kubenswrapper[4575]: E1004 05:54:02.426473 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerName="registry-server" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.426489 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerName="registry-server" Oct 04 05:54:02 crc kubenswrapper[4575]: E1004 05:54:02.426506 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerName="extract-content" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.426513 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerName="extract-content" Oct 04 05:54:02 crc kubenswrapper[4575]: E1004 05:54:02.426545 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerName="extract-utilities" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.426552 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerName="extract-utilities" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.426816 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dedc78d-f8c9-468e-bec8-b0b16d93161c" containerName="registry-server" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.428332 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.443931 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljvqv"] Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.586011 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvlpv\" (UniqueName: \"kubernetes.io/projected/239553b0-6c5d-48f2-ad95-b2f6bfecb327-kube-api-access-mvlpv\") pod \"redhat-marketplace-ljvqv\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.586091 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-catalog-content\") pod \"redhat-marketplace-ljvqv\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.586139 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-utilities\") pod \"redhat-marketplace-ljvqv\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.687938 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-catalog-content\") pod \"redhat-marketplace-ljvqv\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.687996 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-utilities\") pod \"redhat-marketplace-ljvqv\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.688127 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvlpv\" (UniqueName: \"kubernetes.io/projected/239553b0-6c5d-48f2-ad95-b2f6bfecb327-kube-api-access-mvlpv\") pod \"redhat-marketplace-ljvqv\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.688567 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-catalog-content\") pod \"redhat-marketplace-ljvqv\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.688792 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-utilities\") pod \"redhat-marketplace-ljvqv\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.711499 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvlpv\" (UniqueName: \"kubernetes.io/projected/239553b0-6c5d-48f2-ad95-b2f6bfecb327-kube-api-access-mvlpv\") pod \"redhat-marketplace-ljvqv\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.750328 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.828315 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lqpcg"] Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.830560 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.840965 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lqpcg"] Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.994195 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-catalog-content\") pod \"redhat-operators-lqpcg\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.994500 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzvkt\" (UniqueName: \"kubernetes.io/projected/2176b429-ad41-45de-8b93-5d5da3e1991c-kube-api-access-pzvkt\") pod \"redhat-operators-lqpcg\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:02 crc kubenswrapper[4575]: I1004 05:54:02.994642 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-utilities\") pod \"redhat-operators-lqpcg\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.096171 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-utilities\") pod \"redhat-operators-lqpcg\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.096398 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-catalog-content\") pod \"redhat-operators-lqpcg\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.096429 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzvkt\" (UniqueName: \"kubernetes.io/projected/2176b429-ad41-45de-8b93-5d5da3e1991c-kube-api-access-pzvkt\") pod \"redhat-operators-lqpcg\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.097277 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-utilities\") pod \"redhat-operators-lqpcg\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.097380 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-catalog-content\") pod \"redhat-operators-lqpcg\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.130909 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzvkt\" (UniqueName: \"kubernetes.io/projected/2176b429-ad41-45de-8b93-5d5da3e1991c-kube-api-access-pzvkt\") pod \"redhat-operators-lqpcg\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.196472 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.392812 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljvqv"] Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.536299 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljvqv" event={"ID":"239553b0-6c5d-48f2-ad95-b2f6bfecb327","Type":"ContainerStarted","Data":"05b3ce590812ad421100c2010e47a902e4499575e0a7db5fefe7b4dcf3b2e3c2"} Oct 04 05:54:03 crc kubenswrapper[4575]: W1004 05:54:03.772165 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2176b429_ad41_45de_8b93_5d5da3e1991c.slice/crio-d3f944d8e33cbef39a2b68c444821ba8604aa583ea73a13ac28d47e42bc08427 WatchSource:0}: Error finding container d3f944d8e33cbef39a2b68c444821ba8604aa583ea73a13ac28d47e42bc08427: Status 404 returned error can't find the container with id d3f944d8e33cbef39a2b68c444821ba8604aa583ea73a13ac28d47e42bc08427 Oct 04 05:54:03 crc kubenswrapper[4575]: I1004 05:54:03.774961 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lqpcg"] Oct 04 05:54:04 crc kubenswrapper[4575]: I1004 05:54:04.547698 4575 generic.go:334] "Generic (PLEG): container finished" podID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerID="36cd7a4a834fb719f733850a106961c798e97f79ee0bdb036eac46629ed635ba" exitCode=0 Oct 04 05:54:04 crc kubenswrapper[4575]: I1004 05:54:04.547812 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljvqv" event={"ID":"239553b0-6c5d-48f2-ad95-b2f6bfecb327","Type":"ContainerDied","Data":"36cd7a4a834fb719f733850a106961c798e97f79ee0bdb036eac46629ed635ba"} Oct 04 05:54:04 crc kubenswrapper[4575]: I1004 05:54:04.549822 4575 generic.go:334] "Generic (PLEG): container finished" podID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerID="48f6cd89a47d25df0157a0f6379e3e3016613ad3856962e0da18bf2b8e5a5cb9" exitCode=0 Oct 04 05:54:04 crc kubenswrapper[4575]: I1004 05:54:04.549861 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqpcg" event={"ID":"2176b429-ad41-45de-8b93-5d5da3e1991c","Type":"ContainerDied","Data":"48f6cd89a47d25df0157a0f6379e3e3016613ad3856962e0da18bf2b8e5a5cb9"} Oct 04 05:54:04 crc kubenswrapper[4575]: I1004 05:54:04.549884 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqpcg" event={"ID":"2176b429-ad41-45de-8b93-5d5da3e1991c","Type":"ContainerStarted","Data":"d3f944d8e33cbef39a2b68c444821ba8604aa583ea73a13ac28d47e42bc08427"} Oct 04 05:54:06 crc kubenswrapper[4575]: I1004 05:54:06.572992 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqpcg" event={"ID":"2176b429-ad41-45de-8b93-5d5da3e1991c","Type":"ContainerStarted","Data":"d740fb6ccb8da50d543df3ba0a722e18a1475b9564e443d99653957605e460a1"} Oct 04 05:54:06 crc kubenswrapper[4575]: I1004 05:54:06.576854 4575 generic.go:334] "Generic (PLEG): container finished" podID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerID="7f2d766deb45b9aad9249f51491cd3059ad17349a4faaac5cefc60d8f3a3ad05" exitCode=0 Oct 04 05:54:06 crc kubenswrapper[4575]: I1004 05:54:06.576908 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljvqv" event={"ID":"239553b0-6c5d-48f2-ad95-b2f6bfecb327","Type":"ContainerDied","Data":"7f2d766deb45b9aad9249f51491cd3059ad17349a4faaac5cefc60d8f3a3ad05"} Oct 04 05:54:07 crc kubenswrapper[4575]: I1004 05:54:07.588753 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljvqv" event={"ID":"239553b0-6c5d-48f2-ad95-b2f6bfecb327","Type":"ContainerStarted","Data":"ab7823804d161b380ef41a819cf9f1d2d8179cc0d541af564eff523a7f78cfb0"} Oct 04 05:54:07 crc kubenswrapper[4575]: I1004 05:54:07.609087 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ljvqv" podStartSLOduration=3.090764574 podStartE2EDuration="5.609069155s" podCreationTimestamp="2025-10-04 05:54:02 +0000 UTC" firstStartedPulling="2025-10-04 05:54:04.549989787 +0000 UTC m=+4795.878548601" lastFinishedPulling="2025-10-04 05:54:07.068294378 +0000 UTC m=+4798.396853182" observedRunningTime="2025-10-04 05:54:07.606182842 +0000 UTC m=+4798.934741676" watchObservedRunningTime="2025-10-04 05:54:07.609069155 +0000 UTC m=+4798.937627979" Oct 04 05:54:10 crc kubenswrapper[4575]: I1004 05:54:10.622318 4575 generic.go:334] "Generic (PLEG): container finished" podID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerID="d740fb6ccb8da50d543df3ba0a722e18a1475b9564e443d99653957605e460a1" exitCode=0 Oct 04 05:54:10 crc kubenswrapper[4575]: I1004 05:54:10.622413 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqpcg" event={"ID":"2176b429-ad41-45de-8b93-5d5da3e1991c","Type":"ContainerDied","Data":"d740fb6ccb8da50d543df3ba0a722e18a1475b9564e443d99653957605e460a1"} Oct 04 05:54:12 crc kubenswrapper[4575]: I1004 05:54:12.657177 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqpcg" event={"ID":"2176b429-ad41-45de-8b93-5d5da3e1991c","Type":"ContainerStarted","Data":"b426a3f4eedb8608782533da09163a7cf3ffa8f155f508ca1ded2c5f12ea6f07"} Oct 04 05:54:12 crc kubenswrapper[4575]: I1004 05:54:12.677856 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lqpcg" podStartSLOduration=3.704348944 podStartE2EDuration="10.67784046s" podCreationTimestamp="2025-10-04 05:54:02 +0000 UTC" firstStartedPulling="2025-10-04 05:54:04.550954805 +0000 UTC m=+4795.879513619" lastFinishedPulling="2025-10-04 05:54:11.524446321 +0000 UTC m=+4802.853005135" observedRunningTime="2025-10-04 05:54:12.674294518 +0000 UTC m=+4804.002853342" watchObservedRunningTime="2025-10-04 05:54:12.67784046 +0000 UTC m=+4804.006399274" Oct 04 05:54:12 crc kubenswrapper[4575]: I1004 05:54:12.751470 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:12 crc kubenswrapper[4575]: I1004 05:54:12.751760 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:12 crc kubenswrapper[4575]: I1004 05:54:12.807426 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:13 crc kubenswrapper[4575]: I1004 05:54:13.197194 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:13 crc kubenswrapper[4575]: I1004 05:54:13.197240 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:13 crc kubenswrapper[4575]: I1004 05:54:13.723747 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:14 crc kubenswrapper[4575]: I1004 05:54:14.238459 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lqpcg" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerName="registry-server" probeResult="failure" output=< Oct 04 05:54:14 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 05:54:14 crc kubenswrapper[4575]: > Oct 04 05:54:14 crc kubenswrapper[4575]: I1004 05:54:14.820173 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljvqv"] Oct 04 05:54:16 crc kubenswrapper[4575]: I1004 05:54:16.689103 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ljvqv" podUID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerName="registry-server" containerID="cri-o://ab7823804d161b380ef41a819cf9f1d2d8179cc0d541af564eff523a7f78cfb0" gracePeriod=2 Oct 04 05:54:17 crc kubenswrapper[4575]: I1004 05:54:17.703782 4575 generic.go:334] "Generic (PLEG): container finished" podID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerID="ab7823804d161b380ef41a819cf9f1d2d8179cc0d541af564eff523a7f78cfb0" exitCode=0 Oct 04 05:54:17 crc kubenswrapper[4575]: I1004 05:54:17.704274 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljvqv" event={"ID":"239553b0-6c5d-48f2-ad95-b2f6bfecb327","Type":"ContainerDied","Data":"ab7823804d161b380ef41a819cf9f1d2d8179cc0d541af564eff523a7f78cfb0"} Oct 04 05:54:17 crc kubenswrapper[4575]: I1004 05:54:17.834742 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:17 crc kubenswrapper[4575]: I1004 05:54:17.911790 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvlpv\" (UniqueName: \"kubernetes.io/projected/239553b0-6c5d-48f2-ad95-b2f6bfecb327-kube-api-access-mvlpv\") pod \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " Oct 04 05:54:17 crc kubenswrapper[4575]: I1004 05:54:17.911854 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-catalog-content\") pod \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " Oct 04 05:54:17 crc kubenswrapper[4575]: I1004 05:54:17.912080 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-utilities\") pod \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\" (UID: \"239553b0-6c5d-48f2-ad95-b2f6bfecb327\") " Oct 04 05:54:17 crc kubenswrapper[4575]: I1004 05:54:17.914318 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-utilities" (OuterVolumeSpecName: "utilities") pod "239553b0-6c5d-48f2-ad95-b2f6bfecb327" (UID: "239553b0-6c5d-48f2-ad95-b2f6bfecb327"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:54:17 crc kubenswrapper[4575]: I1004 05:54:17.929035 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "239553b0-6c5d-48f2-ad95-b2f6bfecb327" (UID: "239553b0-6c5d-48f2-ad95-b2f6bfecb327"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:54:17 crc kubenswrapper[4575]: I1004 05:54:17.938847 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/239553b0-6c5d-48f2-ad95-b2f6bfecb327-kube-api-access-mvlpv" (OuterVolumeSpecName: "kube-api-access-mvlpv") pod "239553b0-6c5d-48f2-ad95-b2f6bfecb327" (UID: "239553b0-6c5d-48f2-ad95-b2f6bfecb327"). InnerVolumeSpecName "kube-api-access-mvlpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.015113 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvlpv\" (UniqueName: \"kubernetes.io/projected/239553b0-6c5d-48f2-ad95-b2f6bfecb327-kube-api-access-mvlpv\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.015148 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.015157 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239553b0-6c5d-48f2-ad95-b2f6bfecb327-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.714538 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljvqv" event={"ID":"239553b0-6c5d-48f2-ad95-b2f6bfecb327","Type":"ContainerDied","Data":"05b3ce590812ad421100c2010e47a902e4499575e0a7db5fefe7b4dcf3b2e3c2"} Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.714629 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ljvqv" Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.714901 4575 scope.go:117] "RemoveContainer" containerID="ab7823804d161b380ef41a819cf9f1d2d8179cc0d541af564eff523a7f78cfb0" Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.738129 4575 scope.go:117] "RemoveContainer" containerID="7f2d766deb45b9aad9249f51491cd3059ad17349a4faaac5cefc60d8f3a3ad05" Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.761213 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljvqv"] Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.765404 4575 scope.go:117] "RemoveContainer" containerID="36cd7a4a834fb719f733850a106961c798e97f79ee0bdb036eac46629ed635ba" Oct 04 05:54:18 crc kubenswrapper[4575]: I1004 05:54:18.773757 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljvqv"] Oct 04 05:54:19 crc kubenswrapper[4575]: I1004 05:54:19.324789 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" path="/var/lib/kubelet/pods/239553b0-6c5d-48f2-ad95-b2f6bfecb327/volumes" Oct 04 05:54:23 crc kubenswrapper[4575]: I1004 05:54:23.241183 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:23 crc kubenswrapper[4575]: I1004 05:54:23.295995 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:23 crc kubenswrapper[4575]: I1004 05:54:23.481284 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lqpcg"] Oct 04 05:54:24 crc kubenswrapper[4575]: I1004 05:54:24.771696 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lqpcg" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerName="registry-server" containerID="cri-o://b426a3f4eedb8608782533da09163a7cf3ffa8f155f508ca1ded2c5f12ea6f07" gracePeriod=2 Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.783892 4575 generic.go:334] "Generic (PLEG): container finished" podID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerID="b426a3f4eedb8608782533da09163a7cf3ffa8f155f508ca1ded2c5f12ea6f07" exitCode=0 Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.784136 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqpcg" event={"ID":"2176b429-ad41-45de-8b93-5d5da3e1991c","Type":"ContainerDied","Data":"b426a3f4eedb8608782533da09163a7cf3ffa8f155f508ca1ded2c5f12ea6f07"} Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.784351 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lqpcg" event={"ID":"2176b429-ad41-45de-8b93-5d5da3e1991c","Type":"ContainerDied","Data":"d3f944d8e33cbef39a2b68c444821ba8604aa583ea73a13ac28d47e42bc08427"} Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.784394 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3f944d8e33cbef39a2b68c444821ba8604aa583ea73a13ac28d47e42bc08427" Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.845705 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.968008 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzvkt\" (UniqueName: \"kubernetes.io/projected/2176b429-ad41-45de-8b93-5d5da3e1991c-kube-api-access-pzvkt\") pod \"2176b429-ad41-45de-8b93-5d5da3e1991c\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.968327 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-utilities\") pod \"2176b429-ad41-45de-8b93-5d5da3e1991c\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.968363 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-catalog-content\") pod \"2176b429-ad41-45de-8b93-5d5da3e1991c\" (UID: \"2176b429-ad41-45de-8b93-5d5da3e1991c\") " Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.969382 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-utilities" (OuterVolumeSpecName: "utilities") pod "2176b429-ad41-45de-8b93-5d5da3e1991c" (UID: "2176b429-ad41-45de-8b93-5d5da3e1991c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:54:25 crc kubenswrapper[4575]: I1004 05:54:25.975747 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2176b429-ad41-45de-8b93-5d5da3e1991c-kube-api-access-pzvkt" (OuterVolumeSpecName: "kube-api-access-pzvkt") pod "2176b429-ad41-45de-8b93-5d5da3e1991c" (UID: "2176b429-ad41-45de-8b93-5d5da3e1991c"). InnerVolumeSpecName "kube-api-access-pzvkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:54:26 crc kubenswrapper[4575]: I1004 05:54:26.060485 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2176b429-ad41-45de-8b93-5d5da3e1991c" (UID: "2176b429-ad41-45de-8b93-5d5da3e1991c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:54:26 crc kubenswrapper[4575]: I1004 05:54:26.070806 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzvkt\" (UniqueName: \"kubernetes.io/projected/2176b429-ad41-45de-8b93-5d5da3e1991c-kube-api-access-pzvkt\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:26 crc kubenswrapper[4575]: I1004 05:54:26.070862 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:26 crc kubenswrapper[4575]: I1004 05:54:26.070871 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2176b429-ad41-45de-8b93-5d5da3e1991c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:26 crc kubenswrapper[4575]: I1004 05:54:26.792070 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lqpcg" Oct 04 05:54:26 crc kubenswrapper[4575]: I1004 05:54:26.832517 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lqpcg"] Oct 04 05:54:26 crc kubenswrapper[4575]: I1004 05:54:26.840629 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lqpcg"] Oct 04 05:54:27 crc kubenswrapper[4575]: I1004 05:54:27.324127 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" path="/var/lib/kubelet/pods/2176b429-ad41-45de-8b93-5d5da3e1991c/volumes" Oct 04 05:54:38 crc kubenswrapper[4575]: I1004 05:54:38.445807 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:54:38 crc kubenswrapper[4575]: I1004 05:54:38.447779 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:55:08 crc kubenswrapper[4575]: I1004 05:55:08.445902 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:55:08 crc kubenswrapper[4575]: I1004 05:55:08.446666 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:55:38 crc kubenswrapper[4575]: I1004 05:55:38.446786 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:55:38 crc kubenswrapper[4575]: I1004 05:55:38.448280 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:55:38 crc kubenswrapper[4575]: I1004 05:55:38.448411 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:55:38 crc kubenswrapper[4575]: I1004 05:55:38.449348 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c52f1c24086c003b61d72bbb37cb4c73370308383293e59bd9c3a69ae525e0d4"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:55:38 crc kubenswrapper[4575]: I1004 05:55:38.449480 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://c52f1c24086c003b61d72bbb37cb4c73370308383293e59bd9c3a69ae525e0d4" gracePeriod=600 Oct 04 05:55:39 crc kubenswrapper[4575]: I1004 05:55:39.388902 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="c52f1c24086c003b61d72bbb37cb4c73370308383293e59bd9c3a69ae525e0d4" exitCode=0 Oct 04 05:55:39 crc kubenswrapper[4575]: I1004 05:55:39.388955 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"c52f1c24086c003b61d72bbb37cb4c73370308383293e59bd9c3a69ae525e0d4"} Oct 04 05:55:39 crc kubenswrapper[4575]: I1004 05:55:39.389443 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed"} Oct 04 05:55:39 crc kubenswrapper[4575]: I1004 05:55:39.389463 4575 scope.go:117] "RemoveContainer" containerID="6c01355e00cd56fb7242c45375ed3202baed41cd1b88912af2f542de3e14165e" Oct 04 05:57:38 crc kubenswrapper[4575]: I1004 05:57:38.446140 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:57:38 crc kubenswrapper[4575]: I1004 05:57:38.446684 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.532471 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d7vd6"] Oct 04 05:58:04 crc kubenswrapper[4575]: E1004 05:58:04.533626 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerName="extract-content" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.533642 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerName="extract-content" Oct 04 05:58:04 crc kubenswrapper[4575]: E1004 05:58:04.533658 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerName="extract-utilities" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.533665 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerName="extract-utilities" Oct 04 05:58:04 crc kubenswrapper[4575]: E1004 05:58:04.533678 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerName="extract-content" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.533683 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerName="extract-content" Oct 04 05:58:04 crc kubenswrapper[4575]: E1004 05:58:04.533705 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerName="registry-server" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.533712 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerName="registry-server" Oct 04 05:58:04 crc kubenswrapper[4575]: E1004 05:58:04.533735 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerName="registry-server" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.533741 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerName="registry-server" Oct 04 05:58:04 crc kubenswrapper[4575]: E1004 05:58:04.533754 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerName="extract-utilities" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.533763 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerName="extract-utilities" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.533964 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="2176b429-ad41-45de-8b93-5d5da3e1991c" containerName="registry-server" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.533994 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="239553b0-6c5d-48f2-ad95-b2f6bfecb327" containerName="registry-server" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.535347 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.582512 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7vd6"] Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.621297 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-utilities\") pod \"community-operators-d7vd6\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.621350 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6nxn\" (UniqueName: \"kubernetes.io/projected/8d379479-dcd9-48dd-a48a-aeae63b59611-kube-api-access-l6nxn\") pod \"community-operators-d7vd6\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.621424 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-catalog-content\") pod \"community-operators-d7vd6\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.723670 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-utilities\") pod \"community-operators-d7vd6\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.723723 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6nxn\" (UniqueName: \"kubernetes.io/projected/8d379479-dcd9-48dd-a48a-aeae63b59611-kube-api-access-l6nxn\") pod \"community-operators-d7vd6\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.723765 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-catalog-content\") pod \"community-operators-d7vd6\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.724258 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-utilities\") pod \"community-operators-d7vd6\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.724420 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-catalog-content\") pod \"community-operators-d7vd6\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.743891 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6nxn\" (UniqueName: \"kubernetes.io/projected/8d379479-dcd9-48dd-a48a-aeae63b59611-kube-api-access-l6nxn\") pod \"community-operators-d7vd6\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:04 crc kubenswrapper[4575]: I1004 05:58:04.856635 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:05 crc kubenswrapper[4575]: I1004 05:58:05.424542 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7vd6"] Oct 04 05:58:05 crc kubenswrapper[4575]: I1004 05:58:05.704601 4575 generic.go:334] "Generic (PLEG): container finished" podID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerID="c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b" exitCode=0 Oct 04 05:58:05 crc kubenswrapper[4575]: I1004 05:58:05.704823 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7vd6" event={"ID":"8d379479-dcd9-48dd-a48a-aeae63b59611","Type":"ContainerDied","Data":"c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b"} Oct 04 05:58:05 crc kubenswrapper[4575]: I1004 05:58:05.704951 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7vd6" event={"ID":"8d379479-dcd9-48dd-a48a-aeae63b59611","Type":"ContainerStarted","Data":"f3bc6c7fde27c62c644c701b1a27f8584ee1165116618fdbcd29f1e3205fd89f"} Oct 04 05:58:06 crc kubenswrapper[4575]: I1004 05:58:06.715270 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7vd6" event={"ID":"8d379479-dcd9-48dd-a48a-aeae63b59611","Type":"ContainerStarted","Data":"27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0"} Oct 04 05:58:08 crc kubenswrapper[4575]: I1004 05:58:08.446237 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:58:08 crc kubenswrapper[4575]: I1004 05:58:08.446684 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:58:08 crc kubenswrapper[4575]: I1004 05:58:08.733473 4575 generic.go:334] "Generic (PLEG): container finished" podID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerID="27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0" exitCode=0 Oct 04 05:58:08 crc kubenswrapper[4575]: I1004 05:58:08.733533 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7vd6" event={"ID":"8d379479-dcd9-48dd-a48a-aeae63b59611","Type":"ContainerDied","Data":"27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0"} Oct 04 05:58:09 crc kubenswrapper[4575]: I1004 05:58:09.744447 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7vd6" event={"ID":"8d379479-dcd9-48dd-a48a-aeae63b59611","Type":"ContainerStarted","Data":"40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e"} Oct 04 05:58:09 crc kubenswrapper[4575]: I1004 05:58:09.769458 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d7vd6" podStartSLOduration=2.316251833 podStartE2EDuration="5.769442001s" podCreationTimestamp="2025-10-04 05:58:04 +0000 UTC" firstStartedPulling="2025-10-04 05:58:05.707673396 +0000 UTC m=+5037.036232210" lastFinishedPulling="2025-10-04 05:58:09.160863564 +0000 UTC m=+5040.489422378" observedRunningTime="2025-10-04 05:58:09.76175141 +0000 UTC m=+5041.090310214" watchObservedRunningTime="2025-10-04 05:58:09.769442001 +0000 UTC m=+5041.098000815" Oct 04 05:58:14 crc kubenswrapper[4575]: I1004 05:58:14.858108 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:14 crc kubenswrapper[4575]: I1004 05:58:14.859284 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:14 crc kubenswrapper[4575]: I1004 05:58:14.926624 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:15 crc kubenswrapper[4575]: I1004 05:58:15.859226 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:15 crc kubenswrapper[4575]: I1004 05:58:15.922265 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d7vd6"] Oct 04 05:58:17 crc kubenswrapper[4575]: I1004 05:58:17.828838 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d7vd6" podUID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerName="registry-server" containerID="cri-o://40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e" gracePeriod=2 Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.358209 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.439008 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-catalog-content\") pod \"8d379479-dcd9-48dd-a48a-aeae63b59611\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.439357 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-utilities\") pod \"8d379479-dcd9-48dd-a48a-aeae63b59611\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.439492 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6nxn\" (UniqueName: \"kubernetes.io/projected/8d379479-dcd9-48dd-a48a-aeae63b59611-kube-api-access-l6nxn\") pod \"8d379479-dcd9-48dd-a48a-aeae63b59611\" (UID: \"8d379479-dcd9-48dd-a48a-aeae63b59611\") " Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.440221 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-utilities" (OuterVolumeSpecName: "utilities") pod "8d379479-dcd9-48dd-a48a-aeae63b59611" (UID: "8d379479-dcd9-48dd-a48a-aeae63b59611"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.447145 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d379479-dcd9-48dd-a48a-aeae63b59611-kube-api-access-l6nxn" (OuterVolumeSpecName: "kube-api-access-l6nxn") pod "8d379479-dcd9-48dd-a48a-aeae63b59611" (UID: "8d379479-dcd9-48dd-a48a-aeae63b59611"). InnerVolumeSpecName "kube-api-access-l6nxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.500896 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d379479-dcd9-48dd-a48a-aeae63b59611" (UID: "8d379479-dcd9-48dd-a48a-aeae63b59611"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.540917 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.540957 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6nxn\" (UniqueName: \"kubernetes.io/projected/8d379479-dcd9-48dd-a48a-aeae63b59611-kube-api-access-l6nxn\") on node \"crc\" DevicePath \"\"" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.540969 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d379479-dcd9-48dd-a48a-aeae63b59611-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.846735 4575 generic.go:334] "Generic (PLEG): container finished" podID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerID="40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e" exitCode=0 Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.846811 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7vd6" event={"ID":"8d379479-dcd9-48dd-a48a-aeae63b59611","Type":"ContainerDied","Data":"40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e"} Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.846856 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7vd6" event={"ID":"8d379479-dcd9-48dd-a48a-aeae63b59611","Type":"ContainerDied","Data":"f3bc6c7fde27c62c644c701b1a27f8584ee1165116618fdbcd29f1e3205fd89f"} Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.846896 4575 scope.go:117] "RemoveContainer" containerID="40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.846919 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7vd6" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.885404 4575 scope.go:117] "RemoveContainer" containerID="27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.892095 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d7vd6"] Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.903360 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d7vd6"] Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.925125 4575 scope.go:117] "RemoveContainer" containerID="c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.975123 4575 scope.go:117] "RemoveContainer" containerID="40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e" Oct 04 05:58:18 crc kubenswrapper[4575]: E1004 05:58:18.975713 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e\": container with ID starting with 40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e not found: ID does not exist" containerID="40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.975766 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e"} err="failed to get container status \"40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e\": rpc error: code = NotFound desc = could not find container \"40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e\": container with ID starting with 40ae9ce8acb5bc2d143e1e523145670d1640213e45fca4ad34094dff86bc8c5e not found: ID does not exist" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.975792 4575 scope.go:117] "RemoveContainer" containerID="27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0" Oct 04 05:58:18 crc kubenswrapper[4575]: E1004 05:58:18.976318 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0\": container with ID starting with 27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0 not found: ID does not exist" containerID="27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.976419 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0"} err="failed to get container status \"27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0\": rpc error: code = NotFound desc = could not find container \"27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0\": container with ID starting with 27bed60baf72fe509cb1f5e06796bd2e1d6b69250f160834e47f2e84cb6642f0 not found: ID does not exist" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.976440 4575 scope.go:117] "RemoveContainer" containerID="c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b" Oct 04 05:58:18 crc kubenswrapper[4575]: E1004 05:58:18.977034 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b\": container with ID starting with c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b not found: ID does not exist" containerID="c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b" Oct 04 05:58:18 crc kubenswrapper[4575]: I1004 05:58:18.977054 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b"} err="failed to get container status \"c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b\": rpc error: code = NotFound desc = could not find container \"c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b\": container with ID starting with c69ab2a415939d83705142694fe1185dbf586f670d9e23936b394540d750854b not found: ID does not exist" Oct 04 05:58:19 crc kubenswrapper[4575]: I1004 05:58:19.327712 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d379479-dcd9-48dd-a48a-aeae63b59611" path="/var/lib/kubelet/pods/8d379479-dcd9-48dd-a48a-aeae63b59611/volumes" Oct 04 05:58:38 crc kubenswrapper[4575]: I1004 05:58:38.446245 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:58:38 crc kubenswrapper[4575]: I1004 05:58:38.447976 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:58:38 crc kubenswrapper[4575]: I1004 05:58:38.448088 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 05:58:38 crc kubenswrapper[4575]: I1004 05:58:38.448947 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:58:38 crc kubenswrapper[4575]: I1004 05:58:38.449102 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" gracePeriod=600 Oct 04 05:58:38 crc kubenswrapper[4575]: E1004 05:58:38.567433 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:58:39 crc kubenswrapper[4575]: I1004 05:58:39.048321 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" exitCode=0 Oct 04 05:58:39 crc kubenswrapper[4575]: I1004 05:58:39.048426 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed"} Oct 04 05:58:39 crc kubenswrapper[4575]: I1004 05:58:39.048796 4575 scope.go:117] "RemoveContainer" containerID="c52f1c24086c003b61d72bbb37cb4c73370308383293e59bd9c3a69ae525e0d4" Oct 04 05:58:39 crc kubenswrapper[4575]: I1004 05:58:39.049727 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 05:58:39 crc kubenswrapper[4575]: E1004 05:58:39.050072 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:58:51 crc kubenswrapper[4575]: I1004 05:58:51.310469 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 05:58:51 crc kubenswrapper[4575]: E1004 05:58:51.311370 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:59:06 crc kubenswrapper[4575]: I1004 05:59:06.310486 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 05:59:06 crc kubenswrapper[4575]: E1004 05:59:06.311304 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:59:18 crc kubenswrapper[4575]: I1004 05:59:18.310612 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 05:59:18 crc kubenswrapper[4575]: E1004 05:59:18.311419 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:59:33 crc kubenswrapper[4575]: I1004 05:59:33.311575 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 05:59:33 crc kubenswrapper[4575]: E1004 05:59:33.312571 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 05:59:48 crc kubenswrapper[4575]: I1004 05:59:48.310880 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 05:59:48 crc kubenswrapper[4575]: E1004 05:59:48.311600 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.152749 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng"] Oct 04 06:00:00 crc kubenswrapper[4575]: E1004 06:00:00.155970 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerName="extract-utilities" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.156016 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerName="extract-utilities" Oct 04 06:00:00 crc kubenswrapper[4575]: E1004 06:00:00.156029 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerName="registry-server" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.156036 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerName="registry-server" Oct 04 06:00:00 crc kubenswrapper[4575]: E1004 06:00:00.156077 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerName="extract-content" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.156086 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerName="extract-content" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.156373 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d379479-dcd9-48dd-a48a-aeae63b59611" containerName="registry-server" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.157113 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.159787 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.160031 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.169609 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng"] Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.203333 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-config-volume\") pod \"collect-profiles-29325960-pcgng\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.203409 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-secret-volume\") pod \"collect-profiles-29325960-pcgng\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.203465 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hxmb\" (UniqueName: \"kubernetes.io/projected/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-kube-api-access-6hxmb\") pod \"collect-profiles-29325960-pcgng\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.305006 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-secret-volume\") pod \"collect-profiles-29325960-pcgng\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.305072 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hxmb\" (UniqueName: \"kubernetes.io/projected/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-kube-api-access-6hxmb\") pod \"collect-profiles-29325960-pcgng\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.305206 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-config-volume\") pod \"collect-profiles-29325960-pcgng\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.306028 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-config-volume\") pod \"collect-profiles-29325960-pcgng\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.310161 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:00:00 crc kubenswrapper[4575]: E1004 06:00:00.310460 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.314761 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-secret-volume\") pod \"collect-profiles-29325960-pcgng\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.339796 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hxmb\" (UniqueName: \"kubernetes.io/projected/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-kube-api-access-6hxmb\") pod \"collect-profiles-29325960-pcgng\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:00 crc kubenswrapper[4575]: I1004 06:00:00.484933 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:01 crc kubenswrapper[4575]: I1004 06:00:01.012090 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng"] Oct 04 06:00:01 crc kubenswrapper[4575]: I1004 06:00:01.803643 4575 generic.go:334] "Generic (PLEG): container finished" podID="62e63d69-c6af-4d2e-97c0-d6b8491db7a4" containerID="1a45db9ce7a0ac2c8b7d48365ab345d517fdea2faa26a21f370db048b044069e" exitCode=0 Oct 04 06:00:01 crc kubenswrapper[4575]: I1004 06:00:01.803736 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" event={"ID":"62e63d69-c6af-4d2e-97c0-d6b8491db7a4","Type":"ContainerDied","Data":"1a45db9ce7a0ac2c8b7d48365ab345d517fdea2faa26a21f370db048b044069e"} Oct 04 06:00:01 crc kubenswrapper[4575]: I1004 06:00:01.804073 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" event={"ID":"62e63d69-c6af-4d2e-97c0-d6b8491db7a4","Type":"ContainerStarted","Data":"67eb0ebdb41a80437354d954e6b04a5101e5d00003e5f37fcb85a8f21cc6b962"} Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.191855 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.257324 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-config-volume\") pod \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.257418 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hxmb\" (UniqueName: \"kubernetes.io/projected/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-kube-api-access-6hxmb\") pod \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.257611 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-secret-volume\") pod \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\" (UID: \"62e63d69-c6af-4d2e-97c0-d6b8491db7a4\") " Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.258423 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-config-volume" (OuterVolumeSpecName: "config-volume") pod "62e63d69-c6af-4d2e-97c0-d6b8491db7a4" (UID: "62e63d69-c6af-4d2e-97c0-d6b8491db7a4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.263207 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-kube-api-access-6hxmb" (OuterVolumeSpecName: "kube-api-access-6hxmb") pod "62e63d69-c6af-4d2e-97c0-d6b8491db7a4" (UID: "62e63d69-c6af-4d2e-97c0-d6b8491db7a4"). InnerVolumeSpecName "kube-api-access-6hxmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.263852 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "62e63d69-c6af-4d2e-97c0-d6b8491db7a4" (UID: "62e63d69-c6af-4d2e-97c0-d6b8491db7a4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.364222 4575 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.364257 4575 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.364270 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hxmb\" (UniqueName: \"kubernetes.io/projected/62e63d69-c6af-4d2e-97c0-d6b8491db7a4-kube-api-access-6hxmb\") on node \"crc\" DevicePath \"\"" Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.828560 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" event={"ID":"62e63d69-c6af-4d2e-97c0-d6b8491db7a4","Type":"ContainerDied","Data":"67eb0ebdb41a80437354d954e6b04a5101e5d00003e5f37fcb85a8f21cc6b962"} Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.828624 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67eb0ebdb41a80437354d954e6b04a5101e5d00003e5f37fcb85a8f21cc6b962" Oct 04 06:00:03 crc kubenswrapper[4575]: I1004 06:00:03.828648 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-pcgng" Oct 04 06:00:04 crc kubenswrapper[4575]: I1004 06:00:04.262800 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9"] Oct 04 06:00:04 crc kubenswrapper[4575]: I1004 06:00:04.270249 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-mnjm9"] Oct 04 06:00:05 crc kubenswrapper[4575]: I1004 06:00:05.346681 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64a0174c-ba82-4d55-b84c-fe28237cd599" path="/var/lib/kubelet/pods/64a0174c-ba82-4d55-b84c-fe28237cd599/volumes" Oct 04 06:00:15 crc kubenswrapper[4575]: I1004 06:00:15.311259 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:00:15 crc kubenswrapper[4575]: E1004 06:00:15.312110 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:00:28 crc kubenswrapper[4575]: I1004 06:00:28.310043 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:00:28 crc kubenswrapper[4575]: E1004 06:00:28.310848 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:00:32 crc kubenswrapper[4575]: I1004 06:00:32.311254 4575 scope.go:117] "RemoveContainer" containerID="82d2995568e20102a359e4524745e12f70b27edec15f20faf785afb4dabfeae9" Oct 04 06:00:32 crc kubenswrapper[4575]: I1004 06:00:32.349693 4575 scope.go:117] "RemoveContainer" containerID="48f6cd89a47d25df0157a0f6379e3e3016613ad3856962e0da18bf2b8e5a5cb9" Oct 04 06:00:32 crc kubenswrapper[4575]: I1004 06:00:32.437259 4575 scope.go:117] "RemoveContainer" containerID="b426a3f4eedb8608782533da09163a7cf3ffa8f155f508ca1ded2c5f12ea6f07" Oct 04 06:00:32 crc kubenswrapper[4575]: I1004 06:00:32.484872 4575 scope.go:117] "RemoveContainer" containerID="d740fb6ccb8da50d543df3ba0a722e18a1475b9564e443d99653957605e460a1" Oct 04 06:00:43 crc kubenswrapper[4575]: I1004 06:00:43.310940 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:00:43 crc kubenswrapper[4575]: E1004 06:00:43.312072 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:00:55 crc kubenswrapper[4575]: I1004 06:00:55.310958 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:00:55 crc kubenswrapper[4575]: E1004 06:00:55.311936 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.155796 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325961-4dpqf"] Oct 04 06:01:00 crc kubenswrapper[4575]: E1004 06:01:00.157277 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e63d69-c6af-4d2e-97c0-d6b8491db7a4" containerName="collect-profiles" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.157295 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e63d69-c6af-4d2e-97c0-d6b8491db7a4" containerName="collect-profiles" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.157575 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e63d69-c6af-4d2e-97c0-d6b8491db7a4" containerName="collect-profiles" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.158520 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.169991 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325961-4dpqf"] Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.256684 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w8ch\" (UniqueName: \"kubernetes.io/projected/11499d25-b1f8-459e-a57c-f71227734821-kube-api-access-2w8ch\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.256738 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-config-data\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.256762 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-combined-ca-bundle\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.257029 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-fernet-keys\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.359426 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-config-data\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.359531 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-combined-ca-bundle\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.359708 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-fernet-keys\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.359927 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w8ch\" (UniqueName: \"kubernetes.io/projected/11499d25-b1f8-459e-a57c-f71227734821-kube-api-access-2w8ch\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.367827 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-fernet-keys\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.367862 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-config-data\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.368728 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-combined-ca-bundle\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.382507 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w8ch\" (UniqueName: \"kubernetes.io/projected/11499d25-b1f8-459e-a57c-f71227734821-kube-api-access-2w8ch\") pod \"keystone-cron-29325961-4dpqf\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.485611 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:00 crc kubenswrapper[4575]: I1004 06:01:00.953339 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325961-4dpqf"] Oct 04 06:01:00 crc kubenswrapper[4575]: W1004 06:01:00.960948 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11499d25_b1f8_459e_a57c_f71227734821.slice/crio-5ca708069c9a1ee44d4e34ea624262ee5fd1fd9a81cb5cca17482a0553e512b2 WatchSource:0}: Error finding container 5ca708069c9a1ee44d4e34ea624262ee5fd1fd9a81cb5cca17482a0553e512b2: Status 404 returned error can't find the container with id 5ca708069c9a1ee44d4e34ea624262ee5fd1fd9a81cb5cca17482a0553e512b2 Oct 04 06:01:01 crc kubenswrapper[4575]: I1004 06:01:01.413097 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325961-4dpqf" event={"ID":"11499d25-b1f8-459e-a57c-f71227734821","Type":"ContainerStarted","Data":"8988bb717a057c9816a9fd6856370c6dc53cf59e3c76ca93a4678f75d9e17da9"} Oct 04 06:01:01 crc kubenswrapper[4575]: I1004 06:01:01.413173 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325961-4dpqf" event={"ID":"11499d25-b1f8-459e-a57c-f71227734821","Type":"ContainerStarted","Data":"5ca708069c9a1ee44d4e34ea624262ee5fd1fd9a81cb5cca17482a0553e512b2"} Oct 04 06:01:01 crc kubenswrapper[4575]: I1004 06:01:01.440083 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325961-4dpqf" podStartSLOduration=1.440057502 podStartE2EDuration="1.440057502s" podCreationTimestamp="2025-10-04 06:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 06:01:01.433496883 +0000 UTC m=+5212.762055707" watchObservedRunningTime="2025-10-04 06:01:01.440057502 +0000 UTC m=+5212.768616316" Oct 04 06:01:05 crc kubenswrapper[4575]: E1004 06:01:05.836208 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11499d25_b1f8_459e_a57c_f71227734821.slice/crio-8988bb717a057c9816a9fd6856370c6dc53cf59e3c76ca93a4678f75d9e17da9.scope\": RecentStats: unable to find data in memory cache]" Oct 04 06:01:06 crc kubenswrapper[4575]: I1004 06:01:06.469487 4575 generic.go:334] "Generic (PLEG): container finished" podID="11499d25-b1f8-459e-a57c-f71227734821" containerID="8988bb717a057c9816a9fd6856370c6dc53cf59e3c76ca93a4678f75d9e17da9" exitCode=0 Oct 04 06:01:06 crc kubenswrapper[4575]: I1004 06:01:06.469548 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325961-4dpqf" event={"ID":"11499d25-b1f8-459e-a57c-f71227734821","Type":"ContainerDied","Data":"8988bb717a057c9816a9fd6856370c6dc53cf59e3c76ca93a4678f75d9e17da9"} Oct 04 06:01:07 crc kubenswrapper[4575]: I1004 06:01:07.310778 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:01:07 crc kubenswrapper[4575]: E1004 06:01:07.311481 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:01:07 crc kubenswrapper[4575]: I1004 06:01:07.911427 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.060810 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-config-data\") pod \"11499d25-b1f8-459e-a57c-f71227734821\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.060862 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w8ch\" (UniqueName: \"kubernetes.io/projected/11499d25-b1f8-459e-a57c-f71227734821-kube-api-access-2w8ch\") pod \"11499d25-b1f8-459e-a57c-f71227734821\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.060989 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-fernet-keys\") pod \"11499d25-b1f8-459e-a57c-f71227734821\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.061037 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-combined-ca-bundle\") pod \"11499d25-b1f8-459e-a57c-f71227734821\" (UID: \"11499d25-b1f8-459e-a57c-f71227734821\") " Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.067711 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "11499d25-b1f8-459e-a57c-f71227734821" (UID: "11499d25-b1f8-459e-a57c-f71227734821"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.069904 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11499d25-b1f8-459e-a57c-f71227734821-kube-api-access-2w8ch" (OuterVolumeSpecName: "kube-api-access-2w8ch") pod "11499d25-b1f8-459e-a57c-f71227734821" (UID: "11499d25-b1f8-459e-a57c-f71227734821"). InnerVolumeSpecName "kube-api-access-2w8ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.101817 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11499d25-b1f8-459e-a57c-f71227734821" (UID: "11499d25-b1f8-459e-a57c-f71227734821"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.118945 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-config-data" (OuterVolumeSpecName: "config-data") pod "11499d25-b1f8-459e-a57c-f71227734821" (UID: "11499d25-b1f8-459e-a57c-f71227734821"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.165644 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.165691 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w8ch\" (UniqueName: \"kubernetes.io/projected/11499d25-b1f8-459e-a57c-f71227734821-kube-api-access-2w8ch\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.165704 4575 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.165716 4575 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11499d25-b1f8-459e-a57c-f71227734821-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.493761 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325961-4dpqf" event={"ID":"11499d25-b1f8-459e-a57c-f71227734821","Type":"ContainerDied","Data":"5ca708069c9a1ee44d4e34ea624262ee5fd1fd9a81cb5cca17482a0553e512b2"} Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.493812 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ca708069c9a1ee44d4e34ea624262ee5fd1fd9a81cb5cca17482a0553e512b2" Oct 04 06:01:08 crc kubenswrapper[4575]: I1004 06:01:08.493884 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325961-4dpqf" Oct 04 06:01:20 crc kubenswrapper[4575]: I1004 06:01:20.311461 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:01:20 crc kubenswrapper[4575]: E1004 06:01:20.312826 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:01:33 crc kubenswrapper[4575]: I1004 06:01:33.310111 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:01:33 crc kubenswrapper[4575]: E1004 06:01:33.310981 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:01:44 crc kubenswrapper[4575]: I1004 06:01:44.310830 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:01:44 crc kubenswrapper[4575]: E1004 06:01:44.312469 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:01:50 crc kubenswrapper[4575]: I1004 06:01:50.858160 4575 generic.go:334] "Generic (PLEG): container finished" podID="339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" containerID="e52e54d3a67b8e66ee4c34579462f3980bda3212e22973adebbdafe81b7e60c8" exitCode=1 Oct 04 06:01:50 crc kubenswrapper[4575]: I1004 06:01:50.858248 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7","Type":"ContainerDied","Data":"e52e54d3a67b8e66ee4c34579462f3980bda3212e22973adebbdafe81b7e60c8"} Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.311141 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.363277 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ssh-key\") pod \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.363336 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ca-certs\") pod \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.363386 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config\") pod \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.363432 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-config-data\") pod \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.363483 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-workdir\") pod \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.363519 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.363566 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-temporary\") pod \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.369461 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" (UID: "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.372031 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" (UID: "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.374236 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-config-data" (OuterVolumeSpecName: "config-data") pod "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" (UID: "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.376965 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" (UID: "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.398501 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" (UID: "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.404201 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" (UID: "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.416334 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" (UID: "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.465407 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config-secret\") pod \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.465451 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk5w8\" (UniqueName: \"kubernetes.io/projected/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-kube-api-access-fk5w8\") pod \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\" (UID: \"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7\") " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.465710 4575 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.465727 4575 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.465736 4575 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.465746 4575 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.465757 4575 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.465767 4575 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.465787 4575 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.472317 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-kube-api-access-fk5w8" (OuterVolumeSpecName: "kube-api-access-fk5w8") pod "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" (UID: "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7"). InnerVolumeSpecName "kube-api-access-fk5w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.491641 4575 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.506969 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" (UID: "339c2df5-cdf1-4c78-958e-9a4e65e1e2a7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.567123 4575 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.567162 4575 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.567176 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk5w8\" (UniqueName: \"kubernetes.io/projected/339c2df5-cdf1-4c78-958e-9a4e65e1e2a7-kube-api-access-fk5w8\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.878477 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"339c2df5-cdf1-4c78-958e-9a4e65e1e2a7","Type":"ContainerDied","Data":"17fbea7b37bbb9a17ed534d1325cb73c384fc832258ea9b76702cded3d313d06"} Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.878829 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17fbea7b37bbb9a17ed534d1325cb73c384fc832258ea9b76702cded3d313d06" Oct 04 06:01:52 crc kubenswrapper[4575]: I1004 06:01:52.878528 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 06:01:59 crc kubenswrapper[4575]: I1004 06:01:59.321769 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:01:59 crc kubenswrapper[4575]: E1004 06:01:59.322827 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.056283 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 06:02:04 crc kubenswrapper[4575]: E1004 06:02:04.057007 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" containerName="tempest-tests-tempest-tests-runner" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.057021 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" containerName="tempest-tests-tempest-tests-runner" Oct 04 06:02:04 crc kubenswrapper[4575]: E1004 06:02:04.057044 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11499d25-b1f8-459e-a57c-f71227734821" containerName="keystone-cron" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.057050 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="11499d25-b1f8-459e-a57c-f71227734821" containerName="keystone-cron" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.057225 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="339c2df5-cdf1-4c78-958e-9a4e65e1e2a7" containerName="tempest-tests-tempest-tests-runner" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.057251 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="11499d25-b1f8-459e-a57c-f71227734821" containerName="keystone-cron" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.057894 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.061009 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9mw2d" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.074627 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.202905 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d64aafc3-47ee-4084-ba8b-da24a6478ca4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.203239 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htwjm\" (UniqueName: \"kubernetes.io/projected/d64aafc3-47ee-4084-ba8b-da24a6478ca4-kube-api-access-htwjm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d64aafc3-47ee-4084-ba8b-da24a6478ca4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.304944 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htwjm\" (UniqueName: \"kubernetes.io/projected/d64aafc3-47ee-4084-ba8b-da24a6478ca4-kube-api-access-htwjm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d64aafc3-47ee-4084-ba8b-da24a6478ca4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.305058 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d64aafc3-47ee-4084-ba8b-da24a6478ca4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.305518 4575 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d64aafc3-47ee-4084-ba8b-da24a6478ca4\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.327609 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htwjm\" (UniqueName: \"kubernetes.io/projected/d64aafc3-47ee-4084-ba8b-da24a6478ca4-kube-api-access-htwjm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d64aafc3-47ee-4084-ba8b-da24a6478ca4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.335034 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d64aafc3-47ee-4084-ba8b-da24a6478ca4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.392816 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.830918 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.842299 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 06:02:04 crc kubenswrapper[4575]: I1004 06:02:04.997007 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d64aafc3-47ee-4084-ba8b-da24a6478ca4","Type":"ContainerStarted","Data":"b6746eb66cc19b866034944160f8be156d718702bd244122210ac00f4cca5332"} Oct 04 06:02:10 crc kubenswrapper[4575]: I1004 06:02:10.041276 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d64aafc3-47ee-4084-ba8b-da24a6478ca4","Type":"ContainerStarted","Data":"db8bf66ee25c42cb6174213a0e3ef6274df8d6e6860404c1c5d674793fd911a4"} Oct 04 06:02:10 crc kubenswrapper[4575]: I1004 06:02:10.065392 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.035001694 podStartE2EDuration="6.065373141s" podCreationTimestamp="2025-10-04 06:02:04 +0000 UTC" firstStartedPulling="2025-10-04 06:02:04.842028349 +0000 UTC m=+5276.170587173" lastFinishedPulling="2025-10-04 06:02:08.872399806 +0000 UTC m=+5280.200958620" observedRunningTime="2025-10-04 06:02:10.056821554 +0000 UTC m=+5281.385380388" watchObservedRunningTime="2025-10-04 06:02:10.065373141 +0000 UTC m=+5281.393931955" Oct 04 06:02:11 crc kubenswrapper[4575]: I1004 06:02:11.309718 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:02:11 crc kubenswrapper[4575]: E1004 06:02:11.309980 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:02:26 crc kubenswrapper[4575]: I1004 06:02:26.309988 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:02:26 crc kubenswrapper[4575]: E1004 06:02:26.310674 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.624354 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sqpc2/must-gather-l5gpf"] Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.627515 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.629631 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-sqpc2"/"kube-root-ca.crt" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.629878 4575 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-sqpc2"/"openshift-service-ca.crt" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.630126 4575 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-sqpc2"/"default-dockercfg-8klbv" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.651089 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sqpc2/must-gather-l5gpf"] Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.703948 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa422c76-44f8-4120-ba4b-facc21cfbdc5-must-gather-output\") pod \"must-gather-l5gpf\" (UID: \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\") " pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.704183 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2thw\" (UniqueName: \"kubernetes.io/projected/fa422c76-44f8-4120-ba4b-facc21cfbdc5-kube-api-access-r2thw\") pod \"must-gather-l5gpf\" (UID: \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\") " pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.806509 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2thw\" (UniqueName: \"kubernetes.io/projected/fa422c76-44f8-4120-ba4b-facc21cfbdc5-kube-api-access-r2thw\") pod \"must-gather-l5gpf\" (UID: \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\") " pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.806667 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa422c76-44f8-4120-ba4b-facc21cfbdc5-must-gather-output\") pod \"must-gather-l5gpf\" (UID: \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\") " pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.807420 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa422c76-44f8-4120-ba4b-facc21cfbdc5-must-gather-output\") pod \"must-gather-l5gpf\" (UID: \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\") " pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.851616 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2thw\" (UniqueName: \"kubernetes.io/projected/fa422c76-44f8-4120-ba4b-facc21cfbdc5-kube-api-access-r2thw\") pod \"must-gather-l5gpf\" (UID: \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\") " pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:02:35 crc kubenswrapper[4575]: I1004 06:02:35.961032 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:02:36 crc kubenswrapper[4575]: I1004 06:02:36.755063 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sqpc2/must-gather-l5gpf"] Oct 04 06:02:37 crc kubenswrapper[4575]: I1004 06:02:37.296100 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" event={"ID":"fa422c76-44f8-4120-ba4b-facc21cfbdc5","Type":"ContainerStarted","Data":"a92d7b106e4a6e1a70999e9a74ad231d83c83852ab21cd0009d133db01cfe0a2"} Oct 04 06:02:40 crc kubenswrapper[4575]: I1004 06:02:40.310916 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:02:40 crc kubenswrapper[4575]: E1004 06:02:40.311846 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:02:42 crc kubenswrapper[4575]: I1004 06:02:42.376124 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" event={"ID":"fa422c76-44f8-4120-ba4b-facc21cfbdc5","Type":"ContainerStarted","Data":"1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2"} Oct 04 06:02:42 crc kubenswrapper[4575]: I1004 06:02:42.376512 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" event={"ID":"fa422c76-44f8-4120-ba4b-facc21cfbdc5","Type":"ContainerStarted","Data":"4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552"} Oct 04 06:02:47 crc kubenswrapper[4575]: I1004 06:02:47.904893 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" podStartSLOduration=7.999069684 podStartE2EDuration="12.90486883s" podCreationTimestamp="2025-10-04 06:02:35 +0000 UTC" firstStartedPulling="2025-10-04 06:02:36.774304011 +0000 UTC m=+5308.102862825" lastFinishedPulling="2025-10-04 06:02:41.680103157 +0000 UTC m=+5313.008661971" observedRunningTime="2025-10-04 06:02:42.394867885 +0000 UTC m=+5313.723426709" watchObservedRunningTime="2025-10-04 06:02:47.90486883 +0000 UTC m=+5319.233427644" Oct 04 06:02:47 crc kubenswrapper[4575]: I1004 06:02:47.908939 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sqpc2/crc-debug-8mgdk"] Oct 04 06:02:47 crc kubenswrapper[4575]: I1004 06:02:47.910280 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:02:48 crc kubenswrapper[4575]: I1004 06:02:48.020327 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9q7d\" (UniqueName: \"kubernetes.io/projected/dc80465c-3232-4722-8d19-a5b83872e5b2-kube-api-access-j9q7d\") pod \"crc-debug-8mgdk\" (UID: \"dc80465c-3232-4722-8d19-a5b83872e5b2\") " pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:02:48 crc kubenswrapper[4575]: I1004 06:02:48.020448 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc80465c-3232-4722-8d19-a5b83872e5b2-host\") pod \"crc-debug-8mgdk\" (UID: \"dc80465c-3232-4722-8d19-a5b83872e5b2\") " pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:02:48 crc kubenswrapper[4575]: I1004 06:02:48.121712 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc80465c-3232-4722-8d19-a5b83872e5b2-host\") pod \"crc-debug-8mgdk\" (UID: \"dc80465c-3232-4722-8d19-a5b83872e5b2\") " pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:02:48 crc kubenswrapper[4575]: I1004 06:02:48.121869 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc80465c-3232-4722-8d19-a5b83872e5b2-host\") pod \"crc-debug-8mgdk\" (UID: \"dc80465c-3232-4722-8d19-a5b83872e5b2\") " pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:02:48 crc kubenswrapper[4575]: I1004 06:02:48.122201 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9q7d\" (UniqueName: \"kubernetes.io/projected/dc80465c-3232-4722-8d19-a5b83872e5b2-kube-api-access-j9q7d\") pod \"crc-debug-8mgdk\" (UID: \"dc80465c-3232-4722-8d19-a5b83872e5b2\") " pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:02:48 crc kubenswrapper[4575]: I1004 06:02:48.146733 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9q7d\" (UniqueName: \"kubernetes.io/projected/dc80465c-3232-4722-8d19-a5b83872e5b2-kube-api-access-j9q7d\") pod \"crc-debug-8mgdk\" (UID: \"dc80465c-3232-4722-8d19-a5b83872e5b2\") " pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:02:48 crc kubenswrapper[4575]: I1004 06:02:48.236627 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:02:48 crc kubenswrapper[4575]: W1004 06:02:48.276868 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc80465c_3232_4722_8d19_a5b83872e5b2.slice/crio-97e13bd26050fdd244384284a7318cba1c99a8421da25d21f3674646d798db5a WatchSource:0}: Error finding container 97e13bd26050fdd244384284a7318cba1c99a8421da25d21f3674646d798db5a: Status 404 returned error can't find the container with id 97e13bd26050fdd244384284a7318cba1c99a8421da25d21f3674646d798db5a Oct 04 06:02:48 crc kubenswrapper[4575]: I1004 06:02:48.473373 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" event={"ID":"dc80465c-3232-4722-8d19-a5b83872e5b2","Type":"ContainerStarted","Data":"97e13bd26050fdd244384284a7318cba1c99a8421da25d21f3674646d798db5a"} Oct 04 06:02:55 crc kubenswrapper[4575]: I1004 06:02:55.311943 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:02:55 crc kubenswrapper[4575]: E1004 06:02:55.313402 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:02:59 crc kubenswrapper[4575]: I1004 06:02:59.614644 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" event={"ID":"dc80465c-3232-4722-8d19-a5b83872e5b2","Type":"ContainerStarted","Data":"84fa4fa7dea626ee683c1aa9810ed9c961089c7df75d5044efb0b8cb960aed63"} Oct 04 06:02:59 crc kubenswrapper[4575]: I1004 06:02:59.653561 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" podStartSLOduration=2.290621893 podStartE2EDuration="12.653530541s" podCreationTimestamp="2025-10-04 06:02:47 +0000 UTC" firstStartedPulling="2025-10-04 06:02:48.281903531 +0000 UTC m=+5319.610462345" lastFinishedPulling="2025-10-04 06:02:58.644812179 +0000 UTC m=+5329.973370993" observedRunningTime="2025-10-04 06:02:59.646029944 +0000 UTC m=+5330.974588758" watchObservedRunningTime="2025-10-04 06:02:59.653530541 +0000 UTC m=+5330.982089355" Oct 04 06:03:10 crc kubenswrapper[4575]: I1004 06:03:10.310884 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:03:10 crc kubenswrapper[4575]: E1004 06:03:10.311894 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:03:21 crc kubenswrapper[4575]: I1004 06:03:21.313112 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:03:21 crc kubenswrapper[4575]: E1004 06:03:21.314067 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:03:36 crc kubenswrapper[4575]: I1004 06:03:36.310652 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:03:36 crc kubenswrapper[4575]: E1004 06:03:36.311416 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:03:49 crc kubenswrapper[4575]: I1004 06:03:49.323271 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:03:50 crc kubenswrapper[4575]: I1004 06:03:50.185777 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"efcf24d9ad79843003eb9e19166d37813d23fa5a5db13de365f43c79de2bfc62"} Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.609677 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dpmf5"] Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.621800 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.627685 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dpmf5"] Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.758383 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsxsx\" (UniqueName: \"kubernetes.io/projected/69b70667-13ce-49b8-9405-35f01272c2a5-kube-api-access-tsxsx\") pod \"certified-operators-dpmf5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.758501 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-catalog-content\") pod \"certified-operators-dpmf5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.758560 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-utilities\") pod \"certified-operators-dpmf5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.860192 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-catalog-content\") pod \"certified-operators-dpmf5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.860860 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-utilities\") pod \"certified-operators-dpmf5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.860979 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-catalog-content\") pod \"certified-operators-dpmf5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.861162 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsxsx\" (UniqueName: \"kubernetes.io/projected/69b70667-13ce-49b8-9405-35f01272c2a5-kube-api-access-tsxsx\") pod \"certified-operators-dpmf5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.861305 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-utilities\") pod \"certified-operators-dpmf5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.905096 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsxsx\" (UniqueName: \"kubernetes.io/projected/69b70667-13ce-49b8-9405-35f01272c2a5-kube-api-access-tsxsx\") pod \"certified-operators-dpmf5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:58 crc kubenswrapper[4575]: I1004 06:03:58.952830 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:03:59 crc kubenswrapper[4575]: I1004 06:03:59.921996 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dpmf5"] Oct 04 06:04:00 crc kubenswrapper[4575]: I1004 06:04:00.294379 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpmf5" event={"ID":"69b70667-13ce-49b8-9405-35f01272c2a5","Type":"ContainerStarted","Data":"ff7818022267d8cd9709fc62aba05abcebca9a64909f160c95570118440d3625"} Oct 04 06:04:00 crc kubenswrapper[4575]: I1004 06:04:00.294756 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpmf5" event={"ID":"69b70667-13ce-49b8-9405-35f01272c2a5","Type":"ContainerStarted","Data":"e7e8f9fd5bab166fdf2fcfc3c01f414a2a7571aacd925ec8903af851e605fc4d"} Oct 04 06:04:00 crc kubenswrapper[4575]: E1004 06:04:00.512092 4575 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69b70667_13ce_49b8_9405_35f01272c2a5.slice/crio-conmon-ff7818022267d8cd9709fc62aba05abcebca9a64909f160c95570118440d3625.scope\": RecentStats: unable to find data in memory cache]" Oct 04 06:04:01 crc kubenswrapper[4575]: I1004 06:04:01.304858 4575 generic.go:334] "Generic (PLEG): container finished" podID="69b70667-13ce-49b8-9405-35f01272c2a5" containerID="ff7818022267d8cd9709fc62aba05abcebca9a64909f160c95570118440d3625" exitCode=0 Oct 04 06:04:01 crc kubenswrapper[4575]: I1004 06:04:01.305314 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpmf5" event={"ID":"69b70667-13ce-49b8-9405-35f01272c2a5","Type":"ContainerDied","Data":"ff7818022267d8cd9709fc62aba05abcebca9a64909f160c95570118440d3625"} Oct 04 06:04:03 crc kubenswrapper[4575]: I1004 06:04:03.330506 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpmf5" event={"ID":"69b70667-13ce-49b8-9405-35f01272c2a5","Type":"ContainerStarted","Data":"dde0c11f6dff83fdc6cc4de11072d0d7df361f0cea7118eeb52a1ee62fb01db2"} Oct 04 06:04:06 crc kubenswrapper[4575]: I1004 06:04:06.383388 4575 generic.go:334] "Generic (PLEG): container finished" podID="69b70667-13ce-49b8-9405-35f01272c2a5" containerID="dde0c11f6dff83fdc6cc4de11072d0d7df361f0cea7118eeb52a1ee62fb01db2" exitCode=0 Oct 04 06:04:06 crc kubenswrapper[4575]: I1004 06:04:06.383472 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpmf5" event={"ID":"69b70667-13ce-49b8-9405-35f01272c2a5","Type":"ContainerDied","Data":"dde0c11f6dff83fdc6cc4de11072d0d7df361f0cea7118eeb52a1ee62fb01db2"} Oct 04 06:04:07 crc kubenswrapper[4575]: I1004 06:04:07.401332 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpmf5" event={"ID":"69b70667-13ce-49b8-9405-35f01272c2a5","Type":"ContainerStarted","Data":"192576aa5b7e415064d6e29c57c627f75dac58159051532031ff6c106aeb695d"} Oct 04 06:04:07 crc kubenswrapper[4575]: I1004 06:04:07.434756 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dpmf5" podStartSLOduration=3.93718932 podStartE2EDuration="9.434733016s" podCreationTimestamp="2025-10-04 06:03:58 +0000 UTC" firstStartedPulling="2025-10-04 06:04:01.309033379 +0000 UTC m=+5392.637592193" lastFinishedPulling="2025-10-04 06:04:06.806577075 +0000 UTC m=+5398.135135889" observedRunningTime="2025-10-04 06:04:07.426500968 +0000 UTC m=+5398.755059782" watchObservedRunningTime="2025-10-04 06:04:07.434733016 +0000 UTC m=+5398.763291830" Oct 04 06:04:08 crc kubenswrapper[4575]: I1004 06:04:08.953752 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:04:08 crc kubenswrapper[4575]: I1004 06:04:08.955214 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:04:10 crc kubenswrapper[4575]: I1004 06:04:10.015401 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dpmf5" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" containerName="registry-server" probeResult="failure" output=< Oct 04 06:04:10 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 06:04:10 crc kubenswrapper[4575]: > Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.679269 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-59z8r"] Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.683029 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.701167 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-59z8r"] Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.788816 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v5mt\" (UniqueName: \"kubernetes.io/projected/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-kube-api-access-4v5mt\") pod \"redhat-marketplace-59z8r\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.789141 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-catalog-content\") pod \"redhat-marketplace-59z8r\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.789339 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-utilities\") pod \"redhat-marketplace-59z8r\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.891223 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v5mt\" (UniqueName: \"kubernetes.io/projected/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-kube-api-access-4v5mt\") pod \"redhat-marketplace-59z8r\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.891587 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-catalog-content\") pod \"redhat-marketplace-59z8r\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.892112 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-catalog-content\") pod \"redhat-marketplace-59z8r\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.892333 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-utilities\") pod \"redhat-marketplace-59z8r\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.892608 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-utilities\") pod \"redhat-marketplace-59z8r\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:12 crc kubenswrapper[4575]: I1004 06:04:12.914927 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v5mt\" (UniqueName: \"kubernetes.io/projected/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-kube-api-access-4v5mt\") pod \"redhat-marketplace-59z8r\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.004572 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.626884 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-59z8r"] Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.714558 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-txkbl"] Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.719279 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.730724 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-txkbl"] Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.812774 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-utilities\") pod \"redhat-operators-txkbl\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.812840 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-catalog-content\") pod \"redhat-operators-txkbl\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.812924 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff2zh\" (UniqueName: \"kubernetes.io/projected/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-kube-api-access-ff2zh\") pod \"redhat-operators-txkbl\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.915857 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-utilities\") pod \"redhat-operators-txkbl\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.915949 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-catalog-content\") pod \"redhat-operators-txkbl\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.916040 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff2zh\" (UniqueName: \"kubernetes.io/projected/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-kube-api-access-ff2zh\") pod \"redhat-operators-txkbl\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.916457 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-utilities\") pod \"redhat-operators-txkbl\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.916549 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-catalog-content\") pod \"redhat-operators-txkbl\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:13 crc kubenswrapper[4575]: I1004 06:04:13.946045 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff2zh\" (UniqueName: \"kubernetes.io/projected/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-kube-api-access-ff2zh\") pod \"redhat-operators-txkbl\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:14 crc kubenswrapper[4575]: I1004 06:04:14.140280 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:14 crc kubenswrapper[4575]: I1004 06:04:14.482967 4575 generic.go:334] "Generic (PLEG): container finished" podID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerID="ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f" exitCode=0 Oct 04 06:04:14 crc kubenswrapper[4575]: I1004 06:04:14.484557 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-59z8r" event={"ID":"ca7e7fda-ae2b-4025-879b-7b0d4b86c594","Type":"ContainerDied","Data":"ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f"} Oct 04 06:04:14 crc kubenswrapper[4575]: I1004 06:04:14.484600 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-59z8r" event={"ID":"ca7e7fda-ae2b-4025-879b-7b0d4b86c594","Type":"ContainerStarted","Data":"b48b6d915de8abc21893451eb31f3093c8bbb2ada005333ab51447cc4a4e2e3d"} Oct 04 06:04:14 crc kubenswrapper[4575]: I1004 06:04:14.848600 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-txkbl"] Oct 04 06:04:14 crc kubenswrapper[4575]: W1004 06:04:14.870758 4575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0f733f5_ebc1_47c7_91e3_c6c7dba6d9b7.slice/crio-590bb7cf2641b42d98dd7ca6a71c44b91092ea8d1d11deca150faf6f90f59146 WatchSource:0}: Error finding container 590bb7cf2641b42d98dd7ca6a71c44b91092ea8d1d11deca150faf6f90f59146: Status 404 returned error can't find the container with id 590bb7cf2641b42d98dd7ca6a71c44b91092ea8d1d11deca150faf6f90f59146 Oct 04 06:04:15 crc kubenswrapper[4575]: I1004 06:04:15.495209 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-59z8r" event={"ID":"ca7e7fda-ae2b-4025-879b-7b0d4b86c594","Type":"ContainerStarted","Data":"d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2"} Oct 04 06:04:15 crc kubenswrapper[4575]: I1004 06:04:15.497249 4575 generic.go:334] "Generic (PLEG): container finished" podID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerID="67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8" exitCode=0 Oct 04 06:04:15 crc kubenswrapper[4575]: I1004 06:04:15.497306 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txkbl" event={"ID":"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7","Type":"ContainerDied","Data":"67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8"} Oct 04 06:04:15 crc kubenswrapper[4575]: I1004 06:04:15.497334 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txkbl" event={"ID":"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7","Type":"ContainerStarted","Data":"590bb7cf2641b42d98dd7ca6a71c44b91092ea8d1d11deca150faf6f90f59146"} Oct 04 06:04:16 crc kubenswrapper[4575]: I1004 06:04:16.526689 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txkbl" event={"ID":"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7","Type":"ContainerStarted","Data":"1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd"} Oct 04 06:04:16 crc kubenswrapper[4575]: I1004 06:04:16.534391 4575 generic.go:334] "Generic (PLEG): container finished" podID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerID="d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2" exitCode=0 Oct 04 06:04:16 crc kubenswrapper[4575]: I1004 06:04:16.534487 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-59z8r" event={"ID":"ca7e7fda-ae2b-4025-879b-7b0d4b86c594","Type":"ContainerDied","Data":"d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2"} Oct 04 06:04:17 crc kubenswrapper[4575]: I1004 06:04:17.550312 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-59z8r" event={"ID":"ca7e7fda-ae2b-4025-879b-7b0d4b86c594","Type":"ContainerStarted","Data":"1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57"} Oct 04 06:04:17 crc kubenswrapper[4575]: I1004 06:04:17.573376 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-59z8r" podStartSLOduration=2.815307781 podStartE2EDuration="5.573349816s" podCreationTimestamp="2025-10-04 06:04:12 +0000 UTC" firstStartedPulling="2025-10-04 06:04:14.510548994 +0000 UTC m=+5405.839107808" lastFinishedPulling="2025-10-04 06:04:17.268591029 +0000 UTC m=+5408.597149843" observedRunningTime="2025-10-04 06:04:17.572327256 +0000 UTC m=+5408.900886070" watchObservedRunningTime="2025-10-04 06:04:17.573349816 +0000 UTC m=+5408.901908630" Oct 04 06:04:19 crc kubenswrapper[4575]: I1004 06:04:19.030494 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:04:19 crc kubenswrapper[4575]: I1004 06:04:19.102504 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:04:21 crc kubenswrapper[4575]: I1004 06:04:21.868629 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dpmf5"] Oct 04 06:04:21 crc kubenswrapper[4575]: I1004 06:04:21.869126 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dpmf5" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" containerName="registry-server" containerID="cri-o://192576aa5b7e415064d6e29c57c627f75dac58159051532031ff6c106aeb695d" gracePeriod=2 Oct 04 06:04:22 crc kubenswrapper[4575]: I1004 06:04:22.621930 4575 generic.go:334] "Generic (PLEG): container finished" podID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerID="1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd" exitCode=0 Oct 04 06:04:22 crc kubenswrapper[4575]: I1004 06:04:22.622269 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txkbl" event={"ID":"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7","Type":"ContainerDied","Data":"1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd"} Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.005112 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.005180 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.651204 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txkbl" event={"ID":"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7","Type":"ContainerStarted","Data":"a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7"} Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.657180 4575 generic.go:334] "Generic (PLEG): container finished" podID="69b70667-13ce-49b8-9405-35f01272c2a5" containerID="192576aa5b7e415064d6e29c57c627f75dac58159051532031ff6c106aeb695d" exitCode=0 Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.657244 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpmf5" event={"ID":"69b70667-13ce-49b8-9405-35f01272c2a5","Type":"ContainerDied","Data":"192576aa5b7e415064d6e29c57c627f75dac58159051532031ff6c106aeb695d"} Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.657287 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpmf5" event={"ID":"69b70667-13ce-49b8-9405-35f01272c2a5","Type":"ContainerDied","Data":"e7e8f9fd5bab166fdf2fcfc3c01f414a2a7571aacd925ec8903af851e605fc4d"} Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.657301 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7e8f9fd5bab166fdf2fcfc3c01f414a2a7571aacd925ec8903af851e605fc4d" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.691056 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-txkbl" podStartSLOduration=3.127686787 podStartE2EDuration="10.691034433s" podCreationTimestamp="2025-10-04 06:04:13 +0000 UTC" firstStartedPulling="2025-10-04 06:04:15.499444864 +0000 UTC m=+5406.828003678" lastFinishedPulling="2025-10-04 06:04:23.06279251 +0000 UTC m=+5414.391351324" observedRunningTime="2025-10-04 06:04:23.681453637 +0000 UTC m=+5415.010012471" watchObservedRunningTime="2025-10-04 06:04:23.691034433 +0000 UTC m=+5415.019593257" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.694739 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.786822 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-utilities\") pod \"69b70667-13ce-49b8-9405-35f01272c2a5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.786923 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-catalog-content\") pod \"69b70667-13ce-49b8-9405-35f01272c2a5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.787056 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsxsx\" (UniqueName: \"kubernetes.io/projected/69b70667-13ce-49b8-9405-35f01272c2a5-kube-api-access-tsxsx\") pod \"69b70667-13ce-49b8-9405-35f01272c2a5\" (UID: \"69b70667-13ce-49b8-9405-35f01272c2a5\") " Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.787685 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-utilities" (OuterVolumeSpecName: "utilities") pod "69b70667-13ce-49b8-9405-35f01272c2a5" (UID: "69b70667-13ce-49b8-9405-35f01272c2a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.801907 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69b70667-13ce-49b8-9405-35f01272c2a5-kube-api-access-tsxsx" (OuterVolumeSpecName: "kube-api-access-tsxsx") pod "69b70667-13ce-49b8-9405-35f01272c2a5" (UID: "69b70667-13ce-49b8-9405-35f01272c2a5"). InnerVolumeSpecName "kube-api-access-tsxsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.863850 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69b70667-13ce-49b8-9405-35f01272c2a5" (UID: "69b70667-13ce-49b8-9405-35f01272c2a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.890338 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.890380 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b70667-13ce-49b8-9405-35f01272c2a5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:23 crc kubenswrapper[4575]: I1004 06:04:23.890391 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsxsx\" (UniqueName: \"kubernetes.io/projected/69b70667-13ce-49b8-9405-35f01272c2a5-kube-api-access-tsxsx\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:24 crc kubenswrapper[4575]: I1004 06:04:24.070559 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-59z8r" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerName="registry-server" probeResult="failure" output=< Oct 04 06:04:24 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 06:04:24 crc kubenswrapper[4575]: > Oct 04 06:04:24 crc kubenswrapper[4575]: I1004 06:04:24.141263 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:24 crc kubenswrapper[4575]: I1004 06:04:24.141317 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:24 crc kubenswrapper[4575]: I1004 06:04:24.666298 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dpmf5" Oct 04 06:04:24 crc kubenswrapper[4575]: I1004 06:04:24.709858 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dpmf5"] Oct 04 06:04:24 crc kubenswrapper[4575]: I1004 06:04:24.719913 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dpmf5"] Oct 04 06:04:25 crc kubenswrapper[4575]: I1004 06:04:25.209405 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-txkbl" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="registry-server" probeResult="failure" output=< Oct 04 06:04:25 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 06:04:25 crc kubenswrapper[4575]: > Oct 04 06:04:25 crc kubenswrapper[4575]: I1004 06:04:25.323625 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" path="/var/lib/kubelet/pods/69b70667-13ce-49b8-9405-35f01272c2a5/volumes" Oct 04 06:04:33 crc kubenswrapper[4575]: I1004 06:04:33.057360 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:33 crc kubenswrapper[4575]: I1004 06:04:33.113786 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:33 crc kubenswrapper[4575]: I1004 06:04:33.298433 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-59z8r"] Oct 04 06:04:34 crc kubenswrapper[4575]: I1004 06:04:34.755307 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-59z8r" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerName="registry-server" containerID="cri-o://1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57" gracePeriod=2 Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.197458 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-txkbl" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="registry-server" probeResult="failure" output=< Oct 04 06:04:35 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 06:04:35 crc kubenswrapper[4575]: > Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.285060 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.425121 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-utilities\") pod \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.425290 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-catalog-content\") pod \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.425445 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4v5mt\" (UniqueName: \"kubernetes.io/projected/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-kube-api-access-4v5mt\") pod \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\" (UID: \"ca7e7fda-ae2b-4025-879b-7b0d4b86c594\") " Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.426071 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-utilities" (OuterVolumeSpecName: "utilities") pod "ca7e7fda-ae2b-4025-879b-7b0d4b86c594" (UID: "ca7e7fda-ae2b-4025-879b-7b0d4b86c594"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.449938 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca7e7fda-ae2b-4025-879b-7b0d4b86c594" (UID: "ca7e7fda-ae2b-4025-879b-7b0d4b86c594"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.453144 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-kube-api-access-4v5mt" (OuterVolumeSpecName: "kube-api-access-4v5mt") pod "ca7e7fda-ae2b-4025-879b-7b0d4b86c594" (UID: "ca7e7fda-ae2b-4025-879b-7b0d4b86c594"). InnerVolumeSpecName "kube-api-access-4v5mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.527601 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.527636 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4v5mt\" (UniqueName: \"kubernetes.io/projected/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-kube-api-access-4v5mt\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.527646 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7e7fda-ae2b-4025-879b-7b0d4b86c594-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.771402 4575 generic.go:334] "Generic (PLEG): container finished" podID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerID="1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57" exitCode=0 Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.771459 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-59z8r" event={"ID":"ca7e7fda-ae2b-4025-879b-7b0d4b86c594","Type":"ContainerDied","Data":"1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57"} Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.771530 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-59z8r" event={"ID":"ca7e7fda-ae2b-4025-879b-7b0d4b86c594","Type":"ContainerDied","Data":"b48b6d915de8abc21893451eb31f3093c8bbb2ada005333ab51447cc4a4e2e3d"} Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.771533 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-59z8r" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.771554 4575 scope.go:117] "RemoveContainer" containerID="1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.839256 4575 scope.go:117] "RemoveContainer" containerID="d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.859201 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-59z8r"] Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.917223 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-59z8r"] Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.927776 4575 scope.go:117] "RemoveContainer" containerID="ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.955550 4575 scope.go:117] "RemoveContainer" containerID="1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57" Oct 04 06:04:35 crc kubenswrapper[4575]: E1004 06:04:35.958292 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57\": container with ID starting with 1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57 not found: ID does not exist" containerID="1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.958334 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57"} err="failed to get container status \"1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57\": rpc error: code = NotFound desc = could not find container \"1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57\": container with ID starting with 1414b041f9fe2cf03a01668be73123f67ac3ace5fe71db5da3a47648dc29ca57 not found: ID does not exist" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.958359 4575 scope.go:117] "RemoveContainer" containerID="d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2" Oct 04 06:04:35 crc kubenswrapper[4575]: E1004 06:04:35.964377 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2\": container with ID starting with d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2 not found: ID does not exist" containerID="d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.964421 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2"} err="failed to get container status \"d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2\": rpc error: code = NotFound desc = could not find container \"d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2\": container with ID starting with d56f1dc9dac82fc2bf0b832d12a334bb07f116d6ff701b530e134fef9737e0c2 not found: ID does not exist" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.964445 4575 scope.go:117] "RemoveContainer" containerID="ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f" Oct 04 06:04:35 crc kubenswrapper[4575]: E1004 06:04:35.969820 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f\": container with ID starting with ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f not found: ID does not exist" containerID="ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f" Oct 04 06:04:35 crc kubenswrapper[4575]: I1004 06:04:35.969862 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f"} err="failed to get container status \"ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f\": rpc error: code = NotFound desc = could not find container \"ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f\": container with ID starting with ac4c3ce12ffd0a09a5f1e100d264f5b544fd3e4a53eed635b7a525f20ecc358f not found: ID does not exist" Oct 04 06:04:36 crc kubenswrapper[4575]: I1004 06:04:36.780483 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-75d794bf56-cf4hl_3262a5ef-127a-4d3f-ba16-1287e6e4bd9f/barbican-api/0.log" Oct 04 06:04:36 crc kubenswrapper[4575]: I1004 06:04:36.839942 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-75d794bf56-cf4hl_3262a5ef-127a-4d3f-ba16-1287e6e4bd9f/barbican-api-log/0.log" Oct 04 06:04:37 crc kubenswrapper[4575]: I1004 06:04:37.013234 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6d899d7ff8-lv6wl_f83c5777-2ecd-4b0a-a710-4944a1e33ee4/barbican-keystone-listener/0.log" Oct 04 06:04:37 crc kubenswrapper[4575]: I1004 06:04:37.149367 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6d899d7ff8-lv6wl_f83c5777-2ecd-4b0a-a710-4944a1e33ee4/barbican-keystone-listener-log/0.log" Oct 04 06:04:37 crc kubenswrapper[4575]: I1004 06:04:37.294885 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6d6b4c688c-g2g6g_7d9d9151-3e7f-473a-8c53-6ca926cadb6f/barbican-worker/0.log" Oct 04 06:04:37 crc kubenswrapper[4575]: I1004 06:04:37.320881 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" path="/var/lib/kubelet/pods/ca7e7fda-ae2b-4025-879b-7b0d4b86c594/volumes" Oct 04 06:04:37 crc kubenswrapper[4575]: I1004 06:04:37.402274 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6d6b4c688c-g2g6g_7d9d9151-3e7f-473a-8c53-6ca926cadb6f/barbican-worker-log/0.log" Oct 04 06:04:37 crc kubenswrapper[4575]: I1004 06:04:37.671366 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-h4pwp_2c487888-8b86-4aa3-8aaf-19f4ad21f44c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:37 crc kubenswrapper[4575]: I1004 06:04:37.904444 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_36877e92-8c86-4ccd-848c-494f46963048/ceilometer-central-agent/0.log" Oct 04 06:04:37 crc kubenswrapper[4575]: I1004 06:04:37.914450 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_36877e92-8c86-4ccd-848c-494f46963048/proxy-httpd/0.log" Oct 04 06:04:37 crc kubenswrapper[4575]: I1004 06:04:37.985773 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_36877e92-8c86-4ccd-848c-494f46963048/ceilometer-notification-agent/0.log" Oct 04 06:04:38 crc kubenswrapper[4575]: I1004 06:04:38.091356 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_36877e92-8c86-4ccd-848c-494f46963048/sg-core/0.log" Oct 04 06:04:38 crc kubenswrapper[4575]: I1004 06:04:38.262286 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8f959c7b-2fd1-4ee3-8138-8cac2e6342af/cinder-api/0.log" Oct 04 06:04:38 crc kubenswrapper[4575]: I1004 06:04:38.371915 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8f959c7b-2fd1-4ee3-8138-8cac2e6342af/cinder-api-log/0.log" Oct 04 06:04:38 crc kubenswrapper[4575]: I1004 06:04:38.540531 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_da88db0d-f53e-47e2-bdd7-0cd88ca93762/cinder-scheduler/0.log" Oct 04 06:04:38 crc kubenswrapper[4575]: I1004 06:04:38.667447 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_da88db0d-f53e-47e2-bdd7-0cd88ca93762/probe/0.log" Oct 04 06:04:38 crc kubenswrapper[4575]: I1004 06:04:38.845707 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-mrhcd_583857d7-9e34-4be2-9376-b7261f940d9e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:39 crc kubenswrapper[4575]: I1004 06:04:39.095872 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-26wq2_cf592400-cd33-402d-b53f-6e378a7636cc/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:39 crc kubenswrapper[4575]: I1004 06:04:39.196832 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-nmjm7_2fed43b6-f18d-44d6-85f9-c543c5b81935/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:39 crc kubenswrapper[4575]: I1004 06:04:39.335803 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667c9c995c-wgqqr_a830ac92-f57f-4f4d-85e6-55de886289bb/init/0.log" Oct 04 06:04:39 crc kubenswrapper[4575]: I1004 06:04:39.605558 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667c9c995c-wgqqr_a830ac92-f57f-4f4d-85e6-55de886289bb/init/0.log" Oct 04 06:04:39 crc kubenswrapper[4575]: I1004 06:04:39.869238 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667c9c995c-wgqqr_a830ac92-f57f-4f4d-85e6-55de886289bb/dnsmasq-dns/0.log" Oct 04 06:04:40 crc kubenswrapper[4575]: I1004 06:04:40.059630 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-dqsjl_a6d82a60-66ad-4718-ab96-90a5889aeb18/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:40 crc kubenswrapper[4575]: I1004 06:04:40.123445 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9351c479-a1a5-400e-8c8b-358cd9587f8c/glance-httpd/0.log" Oct 04 06:04:40 crc kubenswrapper[4575]: I1004 06:04:40.311948 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9351c479-a1a5-400e-8c8b-358cd9587f8c/glance-log/0.log" Oct 04 06:04:40 crc kubenswrapper[4575]: I1004 06:04:40.452017 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c270011a-523c-473b-ba19-db3a657abe99/glance-httpd/0.log" Oct 04 06:04:40 crc kubenswrapper[4575]: I1004 06:04:40.533433 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c270011a-523c-473b-ba19-db3a657abe99/glance-log/0.log" Oct 04 06:04:40 crc kubenswrapper[4575]: I1004 06:04:40.773381 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-644bf5cdd4-pwdw9_6f0573f0-1418-43eb-be82-166fb99ed23d/horizon/4.log" Oct 04 06:04:40 crc kubenswrapper[4575]: I1004 06:04:40.885375 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-644bf5cdd4-pwdw9_6f0573f0-1418-43eb-be82-166fb99ed23d/horizon/3.log" Oct 04 06:04:41 crc kubenswrapper[4575]: I1004 06:04:41.116499 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-x28j8_24f35162-c12e-4f8b-bce9-798fbf9b38bc/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:41 crc kubenswrapper[4575]: I1004 06:04:41.282969 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-6gfmr_5734782c-1920-445b-9225-f3172b8f28c8/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:41 crc kubenswrapper[4575]: I1004 06:04:41.486186 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-644bf5cdd4-pwdw9_6f0573f0-1418-43eb-be82-166fb99ed23d/horizon-log/0.log" Oct 04 06:04:41 crc kubenswrapper[4575]: I1004 06:04:41.701711 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325901-nx5vx_4c0652b0-f1a2-4439-87b1-2a083fdbb535/keystone-cron/0.log" Oct 04 06:04:42 crc kubenswrapper[4575]: I1004 06:04:42.099348 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-864464787f-5jtr5_04b6e250-1594-4e94-8803-0cf4ed78ebb2/keystone-api/0.log" Oct 04 06:04:42 crc kubenswrapper[4575]: I1004 06:04:42.208085 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325961-4dpqf_11499d25-b1f8-459e-a57c-f71227734821/keystone-cron/0.log" Oct 04 06:04:42 crc kubenswrapper[4575]: I1004 06:04:42.453283 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1a6e2274-1cf7-4762-a0a3-bec545e642c8/kube-state-metrics/0.log" Oct 04 06:04:42 crc kubenswrapper[4575]: I1004 06:04:42.533278 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ns7qg_d143e777-6b76-40a2-886b-a64a66c5d467/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:43 crc kubenswrapper[4575]: I1004 06:04:43.404540 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-656b4b6f57-dw59j_b22cb08b-c72d-4ae3-b161-3e80aca0a636/neutron-httpd/0.log" Oct 04 06:04:43 crc kubenswrapper[4575]: I1004 06:04:43.496887 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-4ptnw_6618ef15-c06c-40a9-85a0-e071b9e13e4f/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:43 crc kubenswrapper[4575]: I1004 06:04:43.583189 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-656b4b6f57-dw59j_b22cb08b-c72d-4ae3-b161-3e80aca0a636/neutron-api/0.log" Oct 04 06:04:43 crc kubenswrapper[4575]: I1004 06:04:43.676206 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3c2dcf5d-de87-4624-bd9a-c1b5a2fdc851/memcached/0.log" Oct 04 06:04:44 crc kubenswrapper[4575]: I1004 06:04:44.452189 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1db1ea6f-24fd-4c11-a2ba-dd47fc6edda2/nova-cell0-conductor-conductor/0.log" Oct 04 06:04:44 crc kubenswrapper[4575]: I1004 06:04:44.773192 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_5823bcd4-8223-4b61-a2a3-aa5cea21fc3b/nova-cell1-conductor-conductor/0.log" Oct 04 06:04:45 crc kubenswrapper[4575]: I1004 06:04:45.190063 4575 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-txkbl" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="registry-server" probeResult="failure" output=< Oct 04 06:04:45 crc kubenswrapper[4575]: timeout: failed to connect service ":50051" within 1s Oct 04 06:04:45 crc kubenswrapper[4575]: > Oct 04 06:04:45 crc kubenswrapper[4575]: I1004 06:04:45.250239 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_8edac101-a91c-4247-a428-53243b6e5c8e/nova-api-log/0.log" Oct 04 06:04:45 crc kubenswrapper[4575]: I1004 06:04:45.257745 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4ddc5e5c-23f0-4bcc-8f6b-32312096a689/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 06:04:45 crc kubenswrapper[4575]: I1004 06:04:45.460896 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-ft442_fde6c2d7-f4e5-4239-9933-5b2f9f4b8a17/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:45 crc kubenswrapper[4575]: I1004 06:04:45.471345 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_8edac101-a91c-4247-a428-53243b6e5c8e/nova-api-api/0.log" Oct 04 06:04:45 crc kubenswrapper[4575]: I1004 06:04:45.587401 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7664c119-13b3-49a5-bd83-ba72b8d54c07/nova-metadata-log/0.log" Oct 04 06:04:46 crc kubenswrapper[4575]: I1004 06:04:46.269846 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_fe22583c-fe25-4a83-a80b-23670a8cf79a/nova-scheduler-scheduler/0.log" Oct 04 06:04:46 crc kubenswrapper[4575]: I1004 06:04:46.334488 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_04440b88-f2b8-418c-9676-714f0e8b7112/mysql-bootstrap/0.log" Oct 04 06:04:46 crc kubenswrapper[4575]: I1004 06:04:46.423678 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_04440b88-f2b8-418c-9676-714f0e8b7112/mysql-bootstrap/0.log" Oct 04 06:04:46 crc kubenswrapper[4575]: I1004 06:04:46.598347 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_04440b88-f2b8-418c-9676-714f0e8b7112/galera/0.log" Oct 04 06:04:46 crc kubenswrapper[4575]: I1004 06:04:46.753075 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_21199967-fd4f-43af-acd4-653828f2335a/mysql-bootstrap/0.log" Oct 04 06:04:47 crc kubenswrapper[4575]: I1004 06:04:47.127997 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_21199967-fd4f-43af-acd4-653828f2335a/mysql-bootstrap/0.log" Oct 04 06:04:47 crc kubenswrapper[4575]: I1004 06:04:47.157160 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_21199967-fd4f-43af-acd4-653828f2335a/galera/0.log" Oct 04 06:04:47 crc kubenswrapper[4575]: I1004 06:04:47.314199 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7664c119-13b3-49a5-bd83-ba72b8d54c07/nova-metadata-metadata/0.log" Oct 04 06:04:47 crc kubenswrapper[4575]: I1004 06:04:47.411410 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f4b579e2-4a74-4589-b2c1-37437f91b776/openstackclient/0.log" Oct 04 06:04:47 crc kubenswrapper[4575]: I1004 06:04:47.510539 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-95h4j_e4e95f59-8c21-495d-bc35-151f36e37681/ovn-controller/0.log" Oct 04 06:04:47 crc kubenswrapper[4575]: I1004 06:04:47.681025 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-xpsnf_ee29b83e-e1a1-4aa2-921a-551c1ff4e3a3/openstack-network-exporter/0.log" Oct 04 06:04:47 crc kubenswrapper[4575]: I1004 06:04:47.813138 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c66wt_1a2a7f61-a13a-4255-83f3-f6c23fc25956/ovsdb-server-init/0.log" Oct 04 06:04:47 crc kubenswrapper[4575]: I1004 06:04:47.976256 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c66wt_1a2a7f61-a13a-4255-83f3-f6c23fc25956/ovsdb-server-init/0.log" Oct 04 06:04:48 crc kubenswrapper[4575]: I1004 06:04:48.023800 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c66wt_1a2a7f61-a13a-4255-83f3-f6c23fc25956/ovs-vswitchd/0.log" Oct 04 06:04:48 crc kubenswrapper[4575]: I1004 06:04:48.056881 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c66wt_1a2a7f61-a13a-4255-83f3-f6c23fc25956/ovsdb-server/0.log" Oct 04 06:04:48 crc kubenswrapper[4575]: I1004 06:04:48.224867 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hdg97_e6092d98-ab7f-4313-86ad-783fa9f625dd/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:48 crc kubenswrapper[4575]: I1004 06:04:48.273668 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2ee38535-11df-486c-a72d-db4a798d4969/openstack-network-exporter/0.log" Oct 04 06:04:48 crc kubenswrapper[4575]: I1004 06:04:48.378152 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2ee38535-11df-486c-a72d-db4a798d4969/ovn-northd/0.log" Oct 04 06:04:48 crc kubenswrapper[4575]: I1004 06:04:48.532442 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bd50f51c-ef8b-4966-a7d5-3f5420295d17/openstack-network-exporter/0.log" Oct 04 06:04:48 crc kubenswrapper[4575]: I1004 06:04:48.539553 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bd50f51c-ef8b-4966-a7d5-3f5420295d17/ovsdbserver-nb/0.log" Oct 04 06:04:48 crc kubenswrapper[4575]: I1004 06:04:48.696088 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ca00752a-2943-4e70-962d-91503f9dd347/openstack-network-exporter/0.log" Oct 04 06:04:48 crc kubenswrapper[4575]: I1004 06:04:48.750399 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ca00752a-2943-4e70-962d-91503f9dd347/ovsdbserver-sb/0.log" Oct 04 06:04:49 crc kubenswrapper[4575]: I1004 06:04:49.107471 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-745885db44-gjnjj_86986b78-6e24-4238-a515-9e5176d6847a/placement-api/0.log" Oct 04 06:04:49 crc kubenswrapper[4575]: I1004 06:04:49.222655 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d699d41f-7e1a-49bc-a88b-1e3906639c7d/setup-container/0.log" Oct 04 06:04:49 crc kubenswrapper[4575]: I1004 06:04:49.232549 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-745885db44-gjnjj_86986b78-6e24-4238-a515-9e5176d6847a/placement-log/0.log" Oct 04 06:04:49 crc kubenswrapper[4575]: I1004 06:04:49.505890 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d699d41f-7e1a-49bc-a88b-1e3906639c7d/rabbitmq/0.log" Oct 04 06:04:49 crc kubenswrapper[4575]: I1004 06:04:49.561525 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d699d41f-7e1a-49bc-a88b-1e3906639c7d/setup-container/0.log" Oct 04 06:04:49 crc kubenswrapper[4575]: I1004 06:04:49.642577 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c02c5e52-debf-4152-8fe1-492d1e88372b/setup-container/0.log" Oct 04 06:04:49 crc kubenswrapper[4575]: I1004 06:04:49.816723 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c02c5e52-debf-4152-8fe1-492d1e88372b/setup-container/0.log" Oct 04 06:04:49 crc kubenswrapper[4575]: I1004 06:04:49.834037 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c02c5e52-debf-4152-8fe1-492d1e88372b/rabbitmq/0.log" Oct 04 06:04:49 crc kubenswrapper[4575]: I1004 06:04:49.949766 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-qdxgd_32a6e686-a532-4a82-9132-4c466a1dc721/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:50 crc kubenswrapper[4575]: I1004 06:04:50.177879 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-25bjt_7a3a20d3-21b8-456a-b46d-bfe264defc14/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:50 crc kubenswrapper[4575]: I1004 06:04:50.252694 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-92l87_727a9553-c9b7-4dab-93fb-c8c9ab1b83a7/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:50 crc kubenswrapper[4575]: I1004 06:04:50.545521 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-rjf7r_f7139366-9c39-49e0-a48f-63dd47e2f52f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:50 crc kubenswrapper[4575]: I1004 06:04:50.601469 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-j4phd_eff0e85f-146c-4e9f-8109-853df39fb0cd/ssh-known-hosts-edpm-deployment/0.log" Oct 04 06:04:50 crc kubenswrapper[4575]: I1004 06:04:50.869615 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-654f8bb69f-d4p2v_b760a20b-b48f-411f-bfb8-34ae6fb9c591/proxy-server/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.032339 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-654f8bb69f-d4p2v_b760a20b-b48f-411f-bfb8-34ae6fb9c591/proxy-httpd/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.111297 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-jw946_d5724856-06a7-4b89-a67e-8239308e798f/swift-ring-rebalance/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.184794 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/account-auditor/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.270885 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/account-reaper/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.413090 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/account-server/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.425233 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/account-replicator/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.470053 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/container-auditor/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.567015 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/container-replicator/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.625079 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/container-server/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.636790 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/container-updater/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.742335 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/object-auditor/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.803715 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/object-expirer/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.852876 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/object-server/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.892981 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/object-replicator/0.log" Oct 04 06:04:51 crc kubenswrapper[4575]: I1004 06:04:51.955772 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/object-updater/0.log" Oct 04 06:04:52 crc kubenswrapper[4575]: I1004 06:04:52.083907 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/swift-recon-cron/0.log" Oct 04 06:04:52 crc kubenswrapper[4575]: I1004 06:04:52.095525 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c9ec5625-8b15-42bd-8258-4960d8000469/rsync/0.log" Oct 04 06:04:52 crc kubenswrapper[4575]: I1004 06:04:52.278010 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-bmlqc_f17c12b6-fe62-4659-ba3f-e61d9a1cac38/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:52 crc kubenswrapper[4575]: I1004 06:04:52.409566 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_339c2df5-cdf1-4c78-958e-9a4e65e1e2a7/tempest-tests-tempest-tests-runner/0.log" Oct 04 06:04:52 crc kubenswrapper[4575]: I1004 06:04:52.586157 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_d64aafc3-47ee-4084-ba8b-da24a6478ca4/test-operator-logs-container/0.log" Oct 04 06:04:52 crc kubenswrapper[4575]: I1004 06:04:52.717974 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-97tnp_89ee882d-998a-4844-a65a-a1a7a1bf9bb7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 06:04:54 crc kubenswrapper[4575]: I1004 06:04:54.205509 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:54 crc kubenswrapper[4575]: I1004 06:04:54.281396 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:54 crc kubenswrapper[4575]: I1004 06:04:54.446509 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-txkbl"] Oct 04 06:04:55 crc kubenswrapper[4575]: I1004 06:04:55.988828 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-txkbl" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="registry-server" containerID="cri-o://a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7" gracePeriod=2 Oct 04 06:04:56 crc kubenswrapper[4575]: I1004 06:04:56.688414 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:56 crc kubenswrapper[4575]: I1004 06:04:56.816876 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-utilities\") pod \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " Oct 04 06:04:56 crc kubenswrapper[4575]: I1004 06:04:56.816946 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-catalog-content\") pod \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " Oct 04 06:04:56 crc kubenswrapper[4575]: I1004 06:04:56.817051 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff2zh\" (UniqueName: \"kubernetes.io/projected/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-kube-api-access-ff2zh\") pod \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\" (UID: \"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7\") " Oct 04 06:04:56 crc kubenswrapper[4575]: I1004 06:04:56.818140 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-utilities" (OuterVolumeSpecName: "utilities") pod "e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" (UID: "e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:04:56 crc kubenswrapper[4575]: I1004 06:04:56.818783 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:56 crc kubenswrapper[4575]: I1004 06:04:56.825128 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-kube-api-access-ff2zh" (OuterVolumeSpecName: "kube-api-access-ff2zh") pod "e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" (UID: "e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7"). InnerVolumeSpecName "kube-api-access-ff2zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:04:56 crc kubenswrapper[4575]: I1004 06:04:56.920516 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff2zh\" (UniqueName: \"kubernetes.io/projected/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-kube-api-access-ff2zh\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:56 crc kubenswrapper[4575]: I1004 06:04:56.923892 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" (UID: "e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.001964 4575 generic.go:334] "Generic (PLEG): container finished" podID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerID="a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7" exitCode=0 Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.002009 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txkbl" event={"ID":"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7","Type":"ContainerDied","Data":"a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7"} Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.002038 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txkbl" event={"ID":"e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7","Type":"ContainerDied","Data":"590bb7cf2641b42d98dd7ca6a71c44b91092ea8d1d11deca150faf6f90f59146"} Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.002055 4575 scope.go:117] "RemoveContainer" containerID="a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.003121 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txkbl" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.022644 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.049670 4575 scope.go:117] "RemoveContainer" containerID="1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.065975 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-txkbl"] Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.074245 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-txkbl"] Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.086548 4575 scope.go:117] "RemoveContainer" containerID="67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.142210 4575 scope.go:117] "RemoveContainer" containerID="a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7" Oct 04 06:04:57 crc kubenswrapper[4575]: E1004 06:04:57.143790 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7\": container with ID starting with a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7 not found: ID does not exist" containerID="a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.143849 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7"} err="failed to get container status \"a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7\": rpc error: code = NotFound desc = could not find container \"a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7\": container with ID starting with a65ad5aea4797ebf64d9489d3d11232659f41b5c05a90a7adf0684c5c5cc06a7 not found: ID does not exist" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.143887 4575 scope.go:117] "RemoveContainer" containerID="1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd" Oct 04 06:04:57 crc kubenswrapper[4575]: E1004 06:04:57.144200 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd\": container with ID starting with 1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd not found: ID does not exist" containerID="1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.144236 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd"} err="failed to get container status \"1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd\": rpc error: code = NotFound desc = could not find container \"1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd\": container with ID starting with 1f661075bc843a16ced73811098cef769f834af13970a79a4d9db1c86974d9dd not found: ID does not exist" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.144258 4575 scope.go:117] "RemoveContainer" containerID="67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8" Oct 04 06:04:57 crc kubenswrapper[4575]: E1004 06:04:57.144547 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8\": container with ID starting with 67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8 not found: ID does not exist" containerID="67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.144615 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8"} err="failed to get container status \"67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8\": rpc error: code = NotFound desc = could not find container \"67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8\": container with ID starting with 67c7e61a11137145be1e54ac88c96088dd7f0e213174ec342439dc4ae1f4bee8 not found: ID does not exist" Oct 04 06:04:57 crc kubenswrapper[4575]: I1004 06:04:57.321511 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" path="/var/lib/kubelet/pods/e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7/volumes" Oct 04 06:05:45 crc kubenswrapper[4575]: I1004 06:05:45.501337 4575 generic.go:334] "Generic (PLEG): container finished" podID="dc80465c-3232-4722-8d19-a5b83872e5b2" containerID="84fa4fa7dea626ee683c1aa9810ed9c961089c7df75d5044efb0b8cb960aed63" exitCode=0 Oct 04 06:05:45 crc kubenswrapper[4575]: I1004 06:05:45.501753 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" event={"ID":"dc80465c-3232-4722-8d19-a5b83872e5b2","Type":"ContainerDied","Data":"84fa4fa7dea626ee683c1aa9810ed9c961089c7df75d5044efb0b8cb960aed63"} Oct 04 06:05:46 crc kubenswrapper[4575]: I1004 06:05:46.623460 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:05:46 crc kubenswrapper[4575]: I1004 06:05:46.652547 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sqpc2/crc-debug-8mgdk"] Oct 04 06:05:46 crc kubenswrapper[4575]: I1004 06:05:46.659914 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sqpc2/crc-debug-8mgdk"] Oct 04 06:05:46 crc kubenswrapper[4575]: I1004 06:05:46.702296 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9q7d\" (UniqueName: \"kubernetes.io/projected/dc80465c-3232-4722-8d19-a5b83872e5b2-kube-api-access-j9q7d\") pod \"dc80465c-3232-4722-8d19-a5b83872e5b2\" (UID: \"dc80465c-3232-4722-8d19-a5b83872e5b2\") " Oct 04 06:05:46 crc kubenswrapper[4575]: I1004 06:05:46.702554 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc80465c-3232-4722-8d19-a5b83872e5b2-host\") pod \"dc80465c-3232-4722-8d19-a5b83872e5b2\" (UID: \"dc80465c-3232-4722-8d19-a5b83872e5b2\") " Oct 04 06:05:46 crc kubenswrapper[4575]: I1004 06:05:46.703030 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc80465c-3232-4722-8d19-a5b83872e5b2-host" (OuterVolumeSpecName: "host") pod "dc80465c-3232-4722-8d19-a5b83872e5b2" (UID: "dc80465c-3232-4722-8d19-a5b83872e5b2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 06:05:46 crc kubenswrapper[4575]: I1004 06:05:46.710078 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc80465c-3232-4722-8d19-a5b83872e5b2-kube-api-access-j9q7d" (OuterVolumeSpecName: "kube-api-access-j9q7d") pod "dc80465c-3232-4722-8d19-a5b83872e5b2" (UID: "dc80465c-3232-4722-8d19-a5b83872e5b2"). InnerVolumeSpecName "kube-api-access-j9q7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:05:46 crc kubenswrapper[4575]: I1004 06:05:46.805214 4575 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc80465c-3232-4722-8d19-a5b83872e5b2-host\") on node \"crc\" DevicePath \"\"" Oct 04 06:05:46 crc kubenswrapper[4575]: I1004 06:05:46.805546 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9q7d\" (UniqueName: \"kubernetes.io/projected/dc80465c-3232-4722-8d19-a5b83872e5b2-kube-api-access-j9q7d\") on node \"crc\" DevicePath \"\"" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.325218 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc80465c-3232-4722-8d19-a5b83872e5b2" path="/var/lib/kubelet/pods/dc80465c-3232-4722-8d19-a5b83872e5b2/volumes" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.520879 4575 scope.go:117] "RemoveContainer" containerID="84fa4fa7dea626ee683c1aa9810ed9c961089c7df75d5044efb0b8cb960aed63" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.520927 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-8mgdk" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.839449 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sqpc2/crc-debug-c86kb"] Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.839931 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc80465c-3232-4722-8d19-a5b83872e5b2" containerName="container-00" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.839944 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc80465c-3232-4722-8d19-a5b83872e5b2" containerName="container-00" Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.839955 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerName="registry-server" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.839960 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerName="registry-server" Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.839978 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerName="extract-content" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.839983 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerName="extract-content" Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.840001 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" containerName="extract-content" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840006 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" containerName="extract-content" Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.840019 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" containerName="extract-utilities" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840025 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" containerName="extract-utilities" Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.840034 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="extract-content" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840039 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="extract-content" Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.840054 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" containerName="registry-server" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840059 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" containerName="registry-server" Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.840070 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="extract-utilities" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840076 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="extract-utilities" Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.840095 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerName="extract-utilities" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840101 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerName="extract-utilities" Oct 04 06:05:47 crc kubenswrapper[4575]: E1004 06:05:47.840109 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="registry-server" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840115 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="registry-server" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840324 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc80465c-3232-4722-8d19-a5b83872e5b2" containerName="container-00" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840342 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="69b70667-13ce-49b8-9405-35f01272c2a5" containerName="registry-server" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840351 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca7e7fda-ae2b-4025-879b-7b0d4b86c594" containerName="registry-server" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840363 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f733f5-ebc1-47c7-91e3-c6c7dba6d9b7" containerName="registry-server" Oct 04 06:05:47 crc kubenswrapper[4575]: I1004 06:05:47.840983 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.027956 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fldp9\" (UniqueName: \"kubernetes.io/projected/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-kube-api-access-fldp9\") pod \"crc-debug-c86kb\" (UID: \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\") " pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.028102 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-host\") pod \"crc-debug-c86kb\" (UID: \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\") " pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.129847 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fldp9\" (UniqueName: \"kubernetes.io/projected/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-kube-api-access-fldp9\") pod \"crc-debug-c86kb\" (UID: \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\") " pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.129998 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-host\") pod \"crc-debug-c86kb\" (UID: \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\") " pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.130168 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-host\") pod \"crc-debug-c86kb\" (UID: \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\") " pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.148241 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fldp9\" (UniqueName: \"kubernetes.io/projected/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-kube-api-access-fldp9\") pod \"crc-debug-c86kb\" (UID: \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\") " pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.155920 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.537388 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/crc-debug-c86kb" event={"ID":"4b7bfa46-bd36-4ea6-b6da-f40bce893bef","Type":"ContainerStarted","Data":"23788336d4df191196abb84a31ae095a726f72d3c46a33f6c747fb1e8bbbb1df"} Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.537428 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/crc-debug-c86kb" event={"ID":"4b7bfa46-bd36-4ea6-b6da-f40bce893bef","Type":"ContainerStarted","Data":"3ac1d5d44cde2fc25e5d5c234f944bbeabf0f48352fcb4af12ffe534bda89d7c"} Oct 04 06:05:48 crc kubenswrapper[4575]: I1004 06:05:48.552579 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sqpc2/crc-debug-c86kb" podStartSLOduration=1.552562773 podStartE2EDuration="1.552562773s" podCreationTimestamp="2025-10-04 06:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 06:05:48.549463674 +0000 UTC m=+5499.878022498" watchObservedRunningTime="2025-10-04 06:05:48.552562773 +0000 UTC m=+5499.881121587" Oct 04 06:05:49 crc kubenswrapper[4575]: I1004 06:05:49.548336 4575 generic.go:334] "Generic (PLEG): container finished" podID="4b7bfa46-bd36-4ea6-b6da-f40bce893bef" containerID="23788336d4df191196abb84a31ae095a726f72d3c46a33f6c747fb1e8bbbb1df" exitCode=0 Oct 04 06:05:49 crc kubenswrapper[4575]: I1004 06:05:49.548717 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/crc-debug-c86kb" event={"ID":"4b7bfa46-bd36-4ea6-b6da-f40bce893bef","Type":"ContainerDied","Data":"23788336d4df191196abb84a31ae095a726f72d3c46a33f6c747fb1e8bbbb1df"} Oct 04 06:05:50 crc kubenswrapper[4575]: I1004 06:05:50.653085 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:50 crc kubenswrapper[4575]: I1004 06:05:50.778705 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-host\") pod \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\" (UID: \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\") " Oct 04 06:05:50 crc kubenswrapper[4575]: I1004 06:05:50.778808 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fldp9\" (UniqueName: \"kubernetes.io/projected/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-kube-api-access-fldp9\") pod \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\" (UID: \"4b7bfa46-bd36-4ea6-b6da-f40bce893bef\") " Oct 04 06:05:50 crc kubenswrapper[4575]: I1004 06:05:50.778820 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-host" (OuterVolumeSpecName: "host") pod "4b7bfa46-bd36-4ea6-b6da-f40bce893bef" (UID: "4b7bfa46-bd36-4ea6-b6da-f40bce893bef"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 06:05:50 crc kubenswrapper[4575]: I1004 06:05:50.779506 4575 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-host\") on node \"crc\" DevicePath \"\"" Oct 04 06:05:50 crc kubenswrapper[4575]: I1004 06:05:50.796266 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-kube-api-access-fldp9" (OuterVolumeSpecName: "kube-api-access-fldp9") pod "4b7bfa46-bd36-4ea6-b6da-f40bce893bef" (UID: "4b7bfa46-bd36-4ea6-b6da-f40bce893bef"). InnerVolumeSpecName "kube-api-access-fldp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:05:50 crc kubenswrapper[4575]: I1004 06:05:50.881258 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fldp9\" (UniqueName: \"kubernetes.io/projected/4b7bfa46-bd36-4ea6-b6da-f40bce893bef-kube-api-access-fldp9\") on node \"crc\" DevicePath \"\"" Oct 04 06:05:51 crc kubenswrapper[4575]: I1004 06:05:51.566148 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/crc-debug-c86kb" event={"ID":"4b7bfa46-bd36-4ea6-b6da-f40bce893bef","Type":"ContainerDied","Data":"3ac1d5d44cde2fc25e5d5c234f944bbeabf0f48352fcb4af12ffe534bda89d7c"} Oct 04 06:05:51 crc kubenswrapper[4575]: I1004 06:05:51.566415 4575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ac1d5d44cde2fc25e5d5c234f944bbeabf0f48352fcb4af12ffe534bda89d7c" Oct 04 06:05:51 crc kubenswrapper[4575]: I1004 06:05:51.566192 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-c86kb" Oct 04 06:05:57 crc kubenswrapper[4575]: I1004 06:05:57.569446 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sqpc2/crc-debug-c86kb"] Oct 04 06:05:57 crc kubenswrapper[4575]: I1004 06:05:57.578740 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sqpc2/crc-debug-c86kb"] Oct 04 06:05:58 crc kubenswrapper[4575]: I1004 06:05:58.755215 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sqpc2/crc-debug-2tfz8"] Oct 04 06:05:58 crc kubenswrapper[4575]: E1004 06:05:58.755971 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b7bfa46-bd36-4ea6-b6da-f40bce893bef" containerName="container-00" Oct 04 06:05:58 crc kubenswrapper[4575]: I1004 06:05:58.755985 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b7bfa46-bd36-4ea6-b6da-f40bce893bef" containerName="container-00" Oct 04 06:05:58 crc kubenswrapper[4575]: I1004 06:05:58.756213 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b7bfa46-bd36-4ea6-b6da-f40bce893bef" containerName="container-00" Oct 04 06:05:58 crc kubenswrapper[4575]: I1004 06:05:58.756957 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:05:58 crc kubenswrapper[4575]: I1004 06:05:58.928388 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e280f417-2e48-466f-bb4b-648b11a3894e-host\") pod \"crc-debug-2tfz8\" (UID: \"e280f417-2e48-466f-bb4b-648b11a3894e\") " pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:05:58 crc kubenswrapper[4575]: I1004 06:05:58.928529 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck5gw\" (UniqueName: \"kubernetes.io/projected/e280f417-2e48-466f-bb4b-648b11a3894e-kube-api-access-ck5gw\") pod \"crc-debug-2tfz8\" (UID: \"e280f417-2e48-466f-bb4b-648b11a3894e\") " pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.031335 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e280f417-2e48-466f-bb4b-648b11a3894e-host\") pod \"crc-debug-2tfz8\" (UID: \"e280f417-2e48-466f-bb4b-648b11a3894e\") " pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.031512 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck5gw\" (UniqueName: \"kubernetes.io/projected/e280f417-2e48-466f-bb4b-648b11a3894e-kube-api-access-ck5gw\") pod \"crc-debug-2tfz8\" (UID: \"e280f417-2e48-466f-bb4b-648b11a3894e\") " pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.031677 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e280f417-2e48-466f-bb4b-648b11a3894e-host\") pod \"crc-debug-2tfz8\" (UID: \"e280f417-2e48-466f-bb4b-648b11a3894e\") " pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.060200 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck5gw\" (UniqueName: \"kubernetes.io/projected/e280f417-2e48-466f-bb4b-648b11a3894e-kube-api-access-ck5gw\") pod \"crc-debug-2tfz8\" (UID: \"e280f417-2e48-466f-bb4b-648b11a3894e\") " pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.105472 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.328537 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b7bfa46-bd36-4ea6-b6da-f40bce893bef" path="/var/lib/kubelet/pods/4b7bfa46-bd36-4ea6-b6da-f40bce893bef/volumes" Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.644053 4575 generic.go:334] "Generic (PLEG): container finished" podID="e280f417-2e48-466f-bb4b-648b11a3894e" containerID="ed24292d04a2af270fbd4329f081a457f76c759c540932efaad0ca317250c7b0" exitCode=0 Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.644139 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" event={"ID":"e280f417-2e48-466f-bb4b-648b11a3894e","Type":"ContainerDied","Data":"ed24292d04a2af270fbd4329f081a457f76c759c540932efaad0ca317250c7b0"} Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.644169 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" event={"ID":"e280f417-2e48-466f-bb4b-648b11a3894e","Type":"ContainerStarted","Data":"eed1908e4ceceb27b723437e9aa5c1f815ab025c4a9fdb83199386b8b9a34749"} Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.686627 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sqpc2/crc-debug-2tfz8"] Oct 04 06:05:59 crc kubenswrapper[4575]: I1004 06:05:59.695863 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sqpc2/crc-debug-2tfz8"] Oct 04 06:06:00 crc kubenswrapper[4575]: I1004 06:06:00.770060 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:06:00 crc kubenswrapper[4575]: I1004 06:06:00.967946 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e280f417-2e48-466f-bb4b-648b11a3894e-host\") pod \"e280f417-2e48-466f-bb4b-648b11a3894e\" (UID: \"e280f417-2e48-466f-bb4b-648b11a3894e\") " Oct 04 06:06:00 crc kubenswrapper[4575]: I1004 06:06:00.968047 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck5gw\" (UniqueName: \"kubernetes.io/projected/e280f417-2e48-466f-bb4b-648b11a3894e-kube-api-access-ck5gw\") pod \"e280f417-2e48-466f-bb4b-648b11a3894e\" (UID: \"e280f417-2e48-466f-bb4b-648b11a3894e\") " Oct 04 06:06:00 crc kubenswrapper[4575]: I1004 06:06:00.968492 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e280f417-2e48-466f-bb4b-648b11a3894e-host" (OuterVolumeSpecName: "host") pod "e280f417-2e48-466f-bb4b-648b11a3894e" (UID: "e280f417-2e48-466f-bb4b-648b11a3894e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 06:06:00 crc kubenswrapper[4575]: I1004 06:06:00.968762 4575 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e280f417-2e48-466f-bb4b-648b11a3894e-host\") on node \"crc\" DevicePath \"\"" Oct 04 06:06:00 crc kubenswrapper[4575]: I1004 06:06:00.973285 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e280f417-2e48-466f-bb4b-648b11a3894e-kube-api-access-ck5gw" (OuterVolumeSpecName: "kube-api-access-ck5gw") pod "e280f417-2e48-466f-bb4b-648b11a3894e" (UID: "e280f417-2e48-466f-bb4b-648b11a3894e"). InnerVolumeSpecName "kube-api-access-ck5gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.070527 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck5gw\" (UniqueName: \"kubernetes.io/projected/e280f417-2e48-466f-bb4b-648b11a3894e-kube-api-access-ck5gw\") on node \"crc\" DevicePath \"\"" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.321462 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e280f417-2e48-466f-bb4b-648b11a3894e" path="/var/lib/kubelet/pods/e280f417-2e48-466f-bb4b-648b11a3894e/volumes" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.377232 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698_15a5cb3d-ea05-4105-9795-7b6839f93323/util/0.log" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.543600 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698_15a5cb3d-ea05-4105-9795-7b6839f93323/util/0.log" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.593348 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698_15a5cb3d-ea05-4105-9795-7b6839f93323/pull/0.log" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.633877 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698_15a5cb3d-ea05-4105-9795-7b6839f93323/pull/0.log" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.662245 4575 scope.go:117] "RemoveContainer" containerID="ed24292d04a2af270fbd4329f081a457f76c759c540932efaad0ca317250c7b0" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.662643 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/crc-debug-2tfz8" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.859900 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698_15a5cb3d-ea05-4105-9795-7b6839f93323/util/0.log" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.866549 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698_15a5cb3d-ea05-4105-9795-7b6839f93323/pull/0.log" Oct 04 06:06:01 crc kubenswrapper[4575]: I1004 06:06:01.872768 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9cb698_15a5cb3d-ea05-4105-9795-7b6839f93323/extract/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.058105 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-ngwbj_4820b955-13b9-4e2e-a8c6-ec51d1626d36/kube-rbac-proxy/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.146874 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-fwjkl_dfe68fb9-4ab5-4935-ae7a-805851f52b6c/kube-rbac-proxy/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.158993 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-ngwbj_4820b955-13b9-4e2e-a8c6-ec51d1626d36/manager/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.363823 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-x5sbn_51b5294c-5b5b-4689-b5b8-179ecd5cdaf3/kube-rbac-proxy/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.370675 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-fwjkl_dfe68fb9-4ab5-4935-ae7a-805851f52b6c/manager/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.437554 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-x5sbn_51b5294c-5b5b-4689-b5b8-179ecd5cdaf3/manager/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.558915 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-6269n_6854e2d6-7f6c-4bff-9a08-3c91ca5fc895/kube-rbac-proxy/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.718324 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-6269n_6854e2d6-7f6c-4bff-9a08-3c91ca5fc895/manager/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.833538 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-62fpr_295ca73b-c75e-477d-9c47-40c53ef128ea/kube-rbac-proxy/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.878669 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-62fpr_295ca73b-c75e-477d-9c47-40c53ef128ea/manager/0.log" Oct 04 06:06:02 crc kubenswrapper[4575]: I1004 06:06:02.963622 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-58mqf_696349ff-0270-4925-8d69-c7b5dd5cb77b/kube-rbac-proxy/0.log" Oct 04 06:06:03 crc kubenswrapper[4575]: I1004 06:06:03.114027 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-58mqf_696349ff-0270-4925-8d69-c7b5dd5cb77b/manager/0.log" Oct 04 06:06:03 crc kubenswrapper[4575]: I1004 06:06:03.185673 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-kffhh_6f920bb1-79e9-42da-88cc-702e70810570/kube-rbac-proxy/0.log" Oct 04 06:06:03 crc kubenswrapper[4575]: I1004 06:06:03.399443 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-c6lwc_1b4368e7-f41a-43b8-988b-13f5d5388926/kube-rbac-proxy/0.log" Oct 04 06:06:03 crc kubenswrapper[4575]: I1004 06:06:03.410295 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-kffhh_6f920bb1-79e9-42da-88cc-702e70810570/manager/0.log" Oct 04 06:06:03 crc kubenswrapper[4575]: I1004 06:06:03.652185 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-c6lwc_1b4368e7-f41a-43b8-988b-13f5d5388926/manager/0.log" Oct 04 06:06:03 crc kubenswrapper[4575]: I1004 06:06:03.908282 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-xhg6m_579db663-7d67-4686-af00-c2bea2e39e98/kube-rbac-proxy/0.log" Oct 04 06:06:03 crc kubenswrapper[4575]: I1004 06:06:03.930671 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-xhg6m_579db663-7d67-4686-af00-c2bea2e39e98/manager/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.010577 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-pwfxn_2a08356f-3bae-483a-a361-2b2e4b930ba5/kube-rbac-proxy/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.135207 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-pwfxn_2a08356f-3bae-483a-a361-2b2e4b930ba5/manager/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.176831 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-pqssw_2c106031-0561-4193-ba35-36aec489385a/kube-rbac-proxy/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.342997 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-pqssw_2c106031-0561-4193-ba35-36aec489385a/manager/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.496153 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-lrjbx_431117ed-e462-4301-99ca-7a6fef717e89/kube-rbac-proxy/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.520657 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-lrjbx_431117ed-e462-4301-99ca-7a6fef717e89/manager/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.671825 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-n288p_f5d7fd17-de45-4b0b-9066-0ae5c85e3c67/kube-rbac-proxy/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.866602 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-n288p_f5d7fd17-de45-4b0b-9066-0ae5c85e3c67/manager/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.896098 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-nnjv5_6721ebee-1d17-4a50-b450-7b912043621b/kube-rbac-proxy/0.log" Oct 04 06:06:04 crc kubenswrapper[4575]: I1004 06:06:04.975872 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-nnjv5_6721ebee-1d17-4a50-b450-7b912043621b/manager/0.log" Oct 04 06:06:05 crc kubenswrapper[4575]: I1004 06:06:05.143876 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw_b27bf13b-6d49-4f06-ba98-cdcdc0958970/kube-rbac-proxy/0.log" Oct 04 06:06:05 crc kubenswrapper[4575]: I1004 06:06:05.151730 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665c8bkmw_b27bf13b-6d49-4f06-ba98-cdcdc0958970/manager/0.log" Oct 04 06:06:05 crc kubenswrapper[4575]: I1004 06:06:05.294427 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c6db87b77-wsfrs_dbca27ef-6c03-4f7b-8682-c30b02a7fcb7/kube-rbac-proxy/0.log" Oct 04 06:06:05 crc kubenswrapper[4575]: I1004 06:06:05.402196 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6dc5776fd9-l2c9l_59e4afb8-7dcb-4c91-b7ff-aac88158d168/kube-rbac-proxy/0.log" Oct 04 06:06:05 crc kubenswrapper[4575]: I1004 06:06:05.756718 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6dc5776fd9-l2c9l_59e4afb8-7dcb-4c91-b7ff-aac88158d168/operator/0.log" Oct 04 06:06:05 crc kubenswrapper[4575]: I1004 06:06:05.927281 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-d6k87_3080dcb3-ddf0-4da4-956a-e90af6fef1b5/registry-server/0.log" Oct 04 06:06:05 crc kubenswrapper[4575]: I1004 06:06:05.949837 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-kkgrw_2c4bb54c-992e-4341-b480-92aa72f49794/kube-rbac-proxy/0.log" Oct 04 06:06:06 crc kubenswrapper[4575]: I1004 06:06:06.205650 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-kkgrw_2c4bb54c-992e-4341-b480-92aa72f49794/manager/0.log" Oct 04 06:06:06 crc kubenswrapper[4575]: I1004 06:06:06.340631 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-psw66_febb4d47-662e-47e3-ab3b-c78defb6a724/manager/0.log" Oct 04 06:06:06 crc kubenswrapper[4575]: I1004 06:06:06.367065 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-psw66_febb4d47-662e-47e3-ab3b-c78defb6a724/kube-rbac-proxy/0.log" Oct 04 06:06:06 crc kubenswrapper[4575]: I1004 06:06:06.660418 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-k22tr_3e326270-271c-4781-bf2d-cf78cef44891/kube-rbac-proxy/0.log" Oct 04 06:06:06 crc kubenswrapper[4575]: I1004 06:06:06.667860 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-nrhzr_6b2f3b8c-07fa-4a90-81a1-2af4feed6070/operator/0.log" Oct 04 06:06:06 crc kubenswrapper[4575]: I1004 06:06:06.712328 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c6db87b77-wsfrs_dbca27ef-6c03-4f7b-8682-c30b02a7fcb7/manager/0.log" Oct 04 06:06:06 crc kubenswrapper[4575]: I1004 06:06:06.875279 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-k22tr_3e326270-271c-4781-bf2d-cf78cef44891/manager/0.log" Oct 04 06:06:06 crc kubenswrapper[4575]: I1004 06:06:06.896624 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-kdcqh_f2f043c4-ba0c-4464-b8f8-fa5fac00f30a/kube-rbac-proxy/0.log" Oct 04 06:06:06 crc kubenswrapper[4575]: I1004 06:06:06.986011 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-kdcqh_f2f043c4-ba0c-4464-b8f8-fa5fac00f30a/manager/0.log" Oct 04 06:06:07 crc kubenswrapper[4575]: I1004 06:06:07.230264 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-4j5ln_3841a418-64a3-4065-8d0c-9b1493f1b7b0/kube-rbac-proxy/0.log" Oct 04 06:06:07 crc kubenswrapper[4575]: I1004 06:06:07.244423 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-4j5ln_3841a418-64a3-4065-8d0c-9b1493f1b7b0/manager/0.log" Oct 04 06:06:07 crc kubenswrapper[4575]: I1004 06:06:07.261397 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-wsbkv_5a45b0e2-20a2-4d00-a588-0b61682b6bb0/kube-rbac-proxy/0.log" Oct 04 06:06:07 crc kubenswrapper[4575]: I1004 06:06:07.313736 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-wsbkv_5a45b0e2-20a2-4d00-a588-0b61682b6bb0/manager/0.log" Oct 04 06:06:08 crc kubenswrapper[4575]: I1004 06:06:08.447753 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:06:08 crc kubenswrapper[4575]: I1004 06:06:08.447829 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:06:23 crc kubenswrapper[4575]: I1004 06:06:23.217194 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5js5f_2fe0947a-c14b-498c-ba57-c5f8733ae76f/control-plane-machine-set-operator/0.log" Oct 04 06:06:23 crc kubenswrapper[4575]: I1004 06:06:23.372391 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rdr7j_6d607e6e-db54-4e24-aafa-8fa7c16c949b/machine-api-operator/0.log" Oct 04 06:06:23 crc kubenswrapper[4575]: I1004 06:06:23.372823 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rdr7j_6d607e6e-db54-4e24-aafa-8fa7c16c949b/kube-rbac-proxy/0.log" Oct 04 06:06:34 crc kubenswrapper[4575]: I1004 06:06:34.968186 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9zwj6_49a0e02e-c4bb-4b20-afa8-3c7cf160cc2f/cert-manager-controller/0.log" Oct 04 06:06:35 crc kubenswrapper[4575]: I1004 06:06:35.314373 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-j4lff_9eb142f2-2b06-40ea-8472-67052416f1e1/cert-manager-webhook/0.log" Oct 04 06:06:35 crc kubenswrapper[4575]: I1004 06:06:35.317960 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-d8xt7_84f5ce18-246c-4e87-83df-c7e31b1d71ec/cert-manager-cainjector/0.log" Oct 04 06:06:38 crc kubenswrapper[4575]: I1004 06:06:38.446132 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:06:38 crc kubenswrapper[4575]: I1004 06:06:38.446424 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:06:48 crc kubenswrapper[4575]: I1004 06:06:48.184230 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-6tqcx_255da756-dc3e-4ea1-946b-82593d8be138/nmstate-console-plugin/0.log" Oct 04 06:06:48 crc kubenswrapper[4575]: I1004 06:06:48.351987 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-d4cmv_38de0af5-93f8-4b73-aadd-9c2932e4de86/nmstate-handler/0.log" Oct 04 06:06:48 crc kubenswrapper[4575]: I1004 06:06:48.563602 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-5d2mv_0155bc5a-f5fc-4e45-b557-d950e8d95ff0/nmstate-metrics/0.log" Oct 04 06:06:48 crc kubenswrapper[4575]: I1004 06:06:48.573537 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-5d2mv_0155bc5a-f5fc-4e45-b557-d950e8d95ff0/kube-rbac-proxy/0.log" Oct 04 06:06:48 crc kubenswrapper[4575]: I1004 06:06:48.765042 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-69wc7_2a669f16-2ea1-485e-9941-4e0cbc92dff1/nmstate-operator/0.log" Oct 04 06:06:48 crc kubenswrapper[4575]: I1004 06:06:48.772168 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-jhfj5_fc5b01a0-9da5-4b69-b0c0-43f01cf3ab1d/nmstate-webhook/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.194309 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-7ztmk_ba9c8a20-378b-44b0-8db7-78d5ced7671e/kube-rbac-proxy/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.297440 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-7ztmk_ba9c8a20-378b-44b0-8db7-78d5ced7671e/controller/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.454194 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-frr-files/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.610497 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-frr-files/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.637126 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-reloader/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.674837 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-reloader/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.677988 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-metrics/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.901255 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-reloader/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.909696 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-frr-files/0.log" Oct 04 06:07:04 crc kubenswrapper[4575]: I1004 06:07:04.951753 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-metrics/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.007760 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-metrics/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.099568 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-frr-files/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.133388 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-reloader/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.187297 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/cp-metrics/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.278252 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/controller/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.327703 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/frr-metrics/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.420960 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/kube-rbac-proxy/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.643543 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/kube-rbac-proxy-frr/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.645849 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/reloader/0.log" Oct 04 06:07:05 crc kubenswrapper[4575]: I1004 06:07:05.976879 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-q5zvr_ea7f8ebe-6b41-4333-9213-8aa2f6180f73/frr-k8s-webhook-server/0.log" Oct 04 06:07:06 crc kubenswrapper[4575]: I1004 06:07:06.060616 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-657bc4848d-dzvq5_4310b719-4247-4689-991a-dffacf39c9a0/manager/0.log" Oct 04 06:07:06 crc kubenswrapper[4575]: I1004 06:07:06.272241 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6d5f8fc9b4-5t69v_c5cbf36b-ad72-4ae9-adeb-3d1a206a4d8a/webhook-server/0.log" Oct 04 06:07:06 crc kubenswrapper[4575]: I1004 06:07:06.496722 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mm46z_eb053b2d-c0d8-456f-9437-ab93727a1900/kube-rbac-proxy/0.log" Oct 04 06:07:07 crc kubenswrapper[4575]: I1004 06:07:07.125207 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9w4x4_27a10d08-b3eb-4d6d-8815-9db3a56313c5/frr/0.log" Oct 04 06:07:07 crc kubenswrapper[4575]: I1004 06:07:07.363183 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mm46z_eb053b2d-c0d8-456f-9437-ab93727a1900/speaker/0.log" Oct 04 06:07:08 crc kubenswrapper[4575]: I1004 06:07:08.446666 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:07:08 crc kubenswrapper[4575]: I1004 06:07:08.446804 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:07:08 crc kubenswrapper[4575]: I1004 06:07:08.446846 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 06:07:08 crc kubenswrapper[4575]: I1004 06:07:08.447887 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"efcf24d9ad79843003eb9e19166d37813d23fa5a5db13de365f43c79de2bfc62"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 06:07:08 crc kubenswrapper[4575]: I1004 06:07:08.447966 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://efcf24d9ad79843003eb9e19166d37813d23fa5a5db13de365f43c79de2bfc62" gracePeriod=600 Oct 04 06:07:09 crc kubenswrapper[4575]: I1004 06:07:09.365745 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="efcf24d9ad79843003eb9e19166d37813d23fa5a5db13de365f43c79de2bfc62" exitCode=0 Oct 04 06:07:09 crc kubenswrapper[4575]: I1004 06:07:09.365816 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"efcf24d9ad79843003eb9e19166d37813d23fa5a5db13de365f43c79de2bfc62"} Oct 04 06:07:09 crc kubenswrapper[4575]: I1004 06:07:09.366192 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerStarted","Data":"3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143"} Oct 04 06:07:09 crc kubenswrapper[4575]: I1004 06:07:09.366221 4575 scope.go:117] "RemoveContainer" containerID="5c8894d2e501444d6102543f5d67975e9a83910194a3cb6308367f58dffd80ed" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.033273 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll_41fb4729-7b5d-4594-a5a6-890cf32a6ec8/util/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.299732 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll_41fb4729-7b5d-4594-a5a6-890cf32a6ec8/pull/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.371535 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll_41fb4729-7b5d-4594-a5a6-890cf32a6ec8/util/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.381954 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll_41fb4729-7b5d-4594-a5a6-890cf32a6ec8/pull/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.558532 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll_41fb4729-7b5d-4594-a5a6-890cf32a6ec8/util/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.601343 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll_41fb4729-7b5d-4594-a5a6-890cf32a6ec8/pull/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.657962 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24nxll_41fb4729-7b5d-4594-a5a6-890cf32a6ec8/extract/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.793748 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9vz74_1aa15892-e136-4c58-aaf2-d62597704254/extract-utilities/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.950275 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9vz74_1aa15892-e136-4c58-aaf2-d62597704254/extract-utilities/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.988158 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9vz74_1aa15892-e136-4c58-aaf2-d62597704254/extract-content/0.log" Oct 04 06:07:19 crc kubenswrapper[4575]: I1004 06:07:19.993915 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9vz74_1aa15892-e136-4c58-aaf2-d62597704254/extract-content/0.log" Oct 04 06:07:20 crc kubenswrapper[4575]: I1004 06:07:20.136972 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9vz74_1aa15892-e136-4c58-aaf2-d62597704254/extract-content/0.log" Oct 04 06:07:20 crc kubenswrapper[4575]: I1004 06:07:20.214920 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9vz74_1aa15892-e136-4c58-aaf2-d62597704254/extract-utilities/0.log" Oct 04 06:07:20 crc kubenswrapper[4575]: I1004 06:07:20.428324 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gd5s6_20397c5b-8ea1-4ffe-851f-d88f1b228967/extract-utilities/0.log" Oct 04 06:07:20 crc kubenswrapper[4575]: I1004 06:07:20.723226 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gd5s6_20397c5b-8ea1-4ffe-851f-d88f1b228967/extract-utilities/0.log" Oct 04 06:07:20 crc kubenswrapper[4575]: I1004 06:07:20.780921 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gd5s6_20397c5b-8ea1-4ffe-851f-d88f1b228967/extract-content/0.log" Oct 04 06:07:20 crc kubenswrapper[4575]: I1004 06:07:20.835311 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gd5s6_20397c5b-8ea1-4ffe-851f-d88f1b228967/extract-content/0.log" Oct 04 06:07:20 crc kubenswrapper[4575]: I1004 06:07:20.956249 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9vz74_1aa15892-e136-4c58-aaf2-d62597704254/registry-server/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.039055 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gd5s6_20397c5b-8ea1-4ffe-851f-d88f1b228967/extract-utilities/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.118375 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gd5s6_20397c5b-8ea1-4ffe-851f-d88f1b228967/extract-content/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.406833 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx_7de5d540-783e-42d5-89fd-b2a7185681b2/util/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.439795 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gd5s6_20397c5b-8ea1-4ffe-851f-d88f1b228967/registry-server/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.602318 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx_7de5d540-783e-42d5-89fd-b2a7185681b2/util/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.624751 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx_7de5d540-783e-42d5-89fd-b2a7185681b2/pull/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.642192 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx_7de5d540-783e-42d5-89fd-b2a7185681b2/pull/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.817452 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx_7de5d540-783e-42d5-89fd-b2a7185681b2/pull/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.831681 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx_7de5d540-783e-42d5-89fd-b2a7185681b2/util/0.log" Oct 04 06:07:21 crc kubenswrapper[4575]: I1004 06:07:21.870184 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c4pqfx_7de5d540-783e-42d5-89fd-b2a7185681b2/extract/0.log" Oct 04 06:07:22 crc kubenswrapper[4575]: I1004 06:07:22.061332 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7nh7v_1f7d3832-1561-4d4f-a47a-311b5aadef90/marketplace-operator/0.log" Oct 04 06:07:22 crc kubenswrapper[4575]: I1004 06:07:22.111145 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncgzm_9f084f23-241a-45b2-88a9-2ffc29b161b4/extract-utilities/0.log" Oct 04 06:07:22 crc kubenswrapper[4575]: I1004 06:07:22.286006 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncgzm_9f084f23-241a-45b2-88a9-2ffc29b161b4/extract-content/0.log" Oct 04 06:07:22 crc kubenswrapper[4575]: I1004 06:07:22.293479 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncgzm_9f084f23-241a-45b2-88a9-2ffc29b161b4/extract-utilities/0.log" Oct 04 06:07:22 crc kubenswrapper[4575]: I1004 06:07:22.339174 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncgzm_9f084f23-241a-45b2-88a9-2ffc29b161b4/extract-content/0.log" Oct 04 06:07:22 crc kubenswrapper[4575]: I1004 06:07:22.554793 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncgzm_9f084f23-241a-45b2-88a9-2ffc29b161b4/extract-utilities/0.log" Oct 04 06:07:22 crc kubenswrapper[4575]: I1004 06:07:22.640884 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncgzm_9f084f23-241a-45b2-88a9-2ffc29b161b4/extract-content/0.log" Oct 04 06:07:22 crc kubenswrapper[4575]: I1004 06:07:22.842307 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qfp6s_cbaf719e-0879-45d7-a285-42109d7fb978/extract-utilities/0.log" Oct 04 06:07:22 crc kubenswrapper[4575]: I1004 06:07:22.846613 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ncgzm_9f084f23-241a-45b2-88a9-2ffc29b161b4/registry-server/0.log" Oct 04 06:07:23 crc kubenswrapper[4575]: I1004 06:07:23.092767 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qfp6s_cbaf719e-0879-45d7-a285-42109d7fb978/extract-utilities/0.log" Oct 04 06:07:23 crc kubenswrapper[4575]: I1004 06:07:23.126164 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qfp6s_cbaf719e-0879-45d7-a285-42109d7fb978/extract-content/0.log" Oct 04 06:07:23 crc kubenswrapper[4575]: I1004 06:07:23.145965 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qfp6s_cbaf719e-0879-45d7-a285-42109d7fb978/extract-content/0.log" Oct 04 06:07:23 crc kubenswrapper[4575]: I1004 06:07:23.270292 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qfp6s_cbaf719e-0879-45d7-a285-42109d7fb978/extract-utilities/0.log" Oct 04 06:07:23 crc kubenswrapper[4575]: I1004 06:07:23.306531 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qfp6s_cbaf719e-0879-45d7-a285-42109d7fb978/extract-content/0.log" Oct 04 06:07:24 crc kubenswrapper[4575]: I1004 06:07:24.156897 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qfp6s_cbaf719e-0879-45d7-a285-42109d7fb978/registry-server/0.log" Oct 04 06:09:21 crc kubenswrapper[4575]: I1004 06:09:21.926218 4575 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fnkb9"] Oct 04 06:09:21 crc kubenswrapper[4575]: E1004 06:09:21.927275 4575 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e280f417-2e48-466f-bb4b-648b11a3894e" containerName="container-00" Oct 04 06:09:21 crc kubenswrapper[4575]: I1004 06:09:21.927293 4575 state_mem.go:107] "Deleted CPUSet assignment" podUID="e280f417-2e48-466f-bb4b-648b11a3894e" containerName="container-00" Oct 04 06:09:21 crc kubenswrapper[4575]: I1004 06:09:21.927506 4575 memory_manager.go:354] "RemoveStaleState removing state" podUID="e280f417-2e48-466f-bb4b-648b11a3894e" containerName="container-00" Oct 04 06:09:21 crc kubenswrapper[4575]: I1004 06:09:21.928987 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:21 crc kubenswrapper[4575]: I1004 06:09:21.960375 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fnkb9"] Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.076685 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-557qg\" (UniqueName: \"kubernetes.io/projected/91feb50f-f42c-4579-95d7-bb708086753d-kube-api-access-557qg\") pod \"community-operators-fnkb9\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.076762 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-utilities\") pod \"community-operators-fnkb9\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.076904 4575 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-catalog-content\") pod \"community-operators-fnkb9\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.178250 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-catalog-content\") pod \"community-operators-fnkb9\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.178397 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-557qg\" (UniqueName: \"kubernetes.io/projected/91feb50f-f42c-4579-95d7-bb708086753d-kube-api-access-557qg\") pod \"community-operators-fnkb9\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.178422 4575 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-utilities\") pod \"community-operators-fnkb9\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.178688 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-catalog-content\") pod \"community-operators-fnkb9\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.178826 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-utilities\") pod \"community-operators-fnkb9\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.203600 4575 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-557qg\" (UniqueName: \"kubernetes.io/projected/91feb50f-f42c-4579-95d7-bb708086753d-kube-api-access-557qg\") pod \"community-operators-fnkb9\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.264152 4575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:22 crc kubenswrapper[4575]: I1004 06:09:22.939531 4575 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fnkb9"] Oct 04 06:09:23 crc kubenswrapper[4575]: I1004 06:09:23.667243 4575 generic.go:334] "Generic (PLEG): container finished" podID="91feb50f-f42c-4579-95d7-bb708086753d" containerID="338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a" exitCode=0 Oct 04 06:09:23 crc kubenswrapper[4575]: I1004 06:09:23.667463 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fnkb9" event={"ID":"91feb50f-f42c-4579-95d7-bb708086753d","Type":"ContainerDied","Data":"338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a"} Oct 04 06:09:23 crc kubenswrapper[4575]: I1004 06:09:23.667790 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fnkb9" event={"ID":"91feb50f-f42c-4579-95d7-bb708086753d","Type":"ContainerStarted","Data":"d8d9f0830df7d21db58bbefe239d426bbe37bdb169633e4d8246fe2825e0705e"} Oct 04 06:09:23 crc kubenswrapper[4575]: I1004 06:09:23.670969 4575 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 06:09:25 crc kubenswrapper[4575]: I1004 06:09:25.693628 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fnkb9" event={"ID":"91feb50f-f42c-4579-95d7-bb708086753d","Type":"ContainerStarted","Data":"974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d"} Oct 04 06:09:27 crc kubenswrapper[4575]: I1004 06:09:27.718252 4575 generic.go:334] "Generic (PLEG): container finished" podID="91feb50f-f42c-4579-95d7-bb708086753d" containerID="974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d" exitCode=0 Oct 04 06:09:27 crc kubenswrapper[4575]: I1004 06:09:27.718312 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fnkb9" event={"ID":"91feb50f-f42c-4579-95d7-bb708086753d","Type":"ContainerDied","Data":"974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d"} Oct 04 06:09:28 crc kubenswrapper[4575]: I1004 06:09:28.728882 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fnkb9" event={"ID":"91feb50f-f42c-4579-95d7-bb708086753d","Type":"ContainerStarted","Data":"ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326"} Oct 04 06:09:28 crc kubenswrapper[4575]: I1004 06:09:28.753207 4575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fnkb9" podStartSLOduration=3.077697414 podStartE2EDuration="7.753189169s" podCreationTimestamp="2025-10-04 06:09:21 +0000 UTC" firstStartedPulling="2025-10-04 06:09:23.670728313 +0000 UTC m=+5714.999287127" lastFinishedPulling="2025-10-04 06:09:28.346220068 +0000 UTC m=+5719.674778882" observedRunningTime="2025-10-04 06:09:28.749872704 +0000 UTC m=+5720.078431518" watchObservedRunningTime="2025-10-04 06:09:28.753189169 +0000 UTC m=+5720.081747983" Oct 04 06:09:32 crc kubenswrapper[4575]: I1004 06:09:32.264907 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:32 crc kubenswrapper[4575]: I1004 06:09:32.265544 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:32 crc kubenswrapper[4575]: I1004 06:09:32.313395 4575 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:38 crc kubenswrapper[4575]: I1004 06:09:38.446570 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:09:38 crc kubenswrapper[4575]: I1004 06:09:38.447415 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:09:42 crc kubenswrapper[4575]: I1004 06:09:42.315459 4575 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:42 crc kubenswrapper[4575]: I1004 06:09:42.373399 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fnkb9"] Oct 04 06:09:42 crc kubenswrapper[4575]: I1004 06:09:42.856814 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fnkb9" podUID="91feb50f-f42c-4579-95d7-bb708086753d" containerName="registry-server" containerID="cri-o://ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326" gracePeriod=2 Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.311683 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.471610 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-557qg\" (UniqueName: \"kubernetes.io/projected/91feb50f-f42c-4579-95d7-bb708086753d-kube-api-access-557qg\") pod \"91feb50f-f42c-4579-95d7-bb708086753d\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.471681 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-catalog-content\") pod \"91feb50f-f42c-4579-95d7-bb708086753d\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.471764 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-utilities\") pod \"91feb50f-f42c-4579-95d7-bb708086753d\" (UID: \"91feb50f-f42c-4579-95d7-bb708086753d\") " Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.473286 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-utilities" (OuterVolumeSpecName: "utilities") pod "91feb50f-f42c-4579-95d7-bb708086753d" (UID: "91feb50f-f42c-4579-95d7-bb708086753d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.486809 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91feb50f-f42c-4579-95d7-bb708086753d-kube-api-access-557qg" (OuterVolumeSpecName: "kube-api-access-557qg") pod "91feb50f-f42c-4579-95d7-bb708086753d" (UID: "91feb50f-f42c-4579-95d7-bb708086753d"). InnerVolumeSpecName "kube-api-access-557qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.526169 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91feb50f-f42c-4579-95d7-bb708086753d" (UID: "91feb50f-f42c-4579-95d7-bb708086753d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.574352 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-557qg\" (UniqueName: \"kubernetes.io/projected/91feb50f-f42c-4579-95d7-bb708086753d-kube-api-access-557qg\") on node \"crc\" DevicePath \"\"" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.574399 4575 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.574413 4575 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91feb50f-f42c-4579-95d7-bb708086753d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.868969 4575 generic.go:334] "Generic (PLEG): container finished" podID="91feb50f-f42c-4579-95d7-bb708086753d" containerID="ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326" exitCode=0 Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.869020 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fnkb9" event={"ID":"91feb50f-f42c-4579-95d7-bb708086753d","Type":"ContainerDied","Data":"ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326"} Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.869049 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fnkb9" event={"ID":"91feb50f-f42c-4579-95d7-bb708086753d","Type":"ContainerDied","Data":"d8d9f0830df7d21db58bbefe239d426bbe37bdb169633e4d8246fe2825e0705e"} Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.869067 4575 scope.go:117] "RemoveContainer" containerID="ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.869177 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fnkb9" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.907040 4575 scope.go:117] "RemoveContainer" containerID="974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.914522 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fnkb9"] Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.927057 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fnkb9"] Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.929244 4575 scope.go:117] "RemoveContainer" containerID="338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.973095 4575 scope.go:117] "RemoveContainer" containerID="ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326" Oct 04 06:09:43 crc kubenswrapper[4575]: E1004 06:09:43.973651 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326\": container with ID starting with ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326 not found: ID does not exist" containerID="ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.973696 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326"} err="failed to get container status \"ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326\": rpc error: code = NotFound desc = could not find container \"ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326\": container with ID starting with ed38cf69b2512961188387e9dbf48f778cbf59491d0b63a3a6276a345cbbc326 not found: ID does not exist" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.973718 4575 scope.go:117] "RemoveContainer" containerID="974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d" Oct 04 06:09:43 crc kubenswrapper[4575]: E1004 06:09:43.974098 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d\": container with ID starting with 974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d not found: ID does not exist" containerID="974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.974180 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d"} err="failed to get container status \"974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d\": rpc error: code = NotFound desc = could not find container \"974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d\": container with ID starting with 974da959d77f00eaa828b47a5780db91f9f65a1acadcf7e63418f0b27e20e55d not found: ID does not exist" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.974208 4575 scope.go:117] "RemoveContainer" containerID="338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a" Oct 04 06:09:43 crc kubenswrapper[4575]: E1004 06:09:43.974674 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a\": container with ID starting with 338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a not found: ID does not exist" containerID="338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a" Oct 04 06:09:43 crc kubenswrapper[4575]: I1004 06:09:43.974698 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a"} err="failed to get container status \"338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a\": rpc error: code = NotFound desc = could not find container \"338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a\": container with ID starting with 338afb62b94c356d4a3470375fa45dd7f441105f1eb97e14a8ff169c16d86a4a not found: ID does not exist" Oct 04 06:09:45 crc kubenswrapper[4575]: I1004 06:09:45.321057 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91feb50f-f42c-4579-95d7-bb708086753d" path="/var/lib/kubelet/pods/91feb50f-f42c-4579-95d7-bb708086753d/volumes" Oct 04 06:10:08 crc kubenswrapper[4575]: I1004 06:10:08.446310 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:10:08 crc kubenswrapper[4575]: I1004 06:10:08.446879 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:10:32 crc kubenswrapper[4575]: I1004 06:10:32.838787 4575 scope.go:117] "RemoveContainer" containerID="192576aa5b7e415064d6e29c57c627f75dac58159051532031ff6c106aeb695d" Oct 04 06:10:32 crc kubenswrapper[4575]: I1004 06:10:32.860183 4575 scope.go:117] "RemoveContainer" containerID="dde0c11f6dff83fdc6cc4de11072d0d7df361f0cea7118eeb52a1ee62fb01db2" Oct 04 06:10:32 crc kubenswrapper[4575]: I1004 06:10:32.883629 4575 scope.go:117] "RemoveContainer" containerID="ff7818022267d8cd9709fc62aba05abcebca9a64909f160c95570118440d3625" Oct 04 06:10:38 crc kubenswrapper[4575]: I1004 06:10:38.446372 4575 patch_prober.go:28] interesting pod/machine-config-daemon-pl4jq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:10:38 crc kubenswrapper[4575]: I1004 06:10:38.447340 4575 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:10:38 crc kubenswrapper[4575]: I1004 06:10:38.447436 4575 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" Oct 04 06:10:38 crc kubenswrapper[4575]: I1004 06:10:38.448849 4575 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143"} pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 06:10:38 crc kubenswrapper[4575]: I1004 06:10:38.448925 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerName="machine-config-daemon" containerID="cri-o://3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" gracePeriod=600 Oct 04 06:10:39 crc kubenswrapper[4575]: E1004 06:10:39.119733 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:10:39 crc kubenswrapper[4575]: I1004 06:10:39.399421 4575 generic.go:334] "Generic (PLEG): container finished" podID="fa422c76-44f8-4120-ba4b-facc21cfbdc5" containerID="4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552" exitCode=0 Oct 04 06:10:39 crc kubenswrapper[4575]: I1004 06:10:39.399460 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" event={"ID":"fa422c76-44f8-4120-ba4b-facc21cfbdc5","Type":"ContainerDied","Data":"4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552"} Oct 04 06:10:39 crc kubenswrapper[4575]: I1004 06:10:39.400082 4575 scope.go:117] "RemoveContainer" containerID="4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552" Oct 04 06:10:39 crc kubenswrapper[4575]: I1004 06:10:39.402372 4575 generic.go:334] "Generic (PLEG): container finished" podID="df9f335e-3c50-4cbf-ad47-491eb2486de8" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" exitCode=0 Oct 04 06:10:39 crc kubenswrapper[4575]: I1004 06:10:39.402406 4575 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" event={"ID":"df9f335e-3c50-4cbf-ad47-491eb2486de8","Type":"ContainerDied","Data":"3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143"} Oct 04 06:10:39 crc kubenswrapper[4575]: I1004 06:10:39.402430 4575 scope.go:117] "RemoveContainer" containerID="efcf24d9ad79843003eb9e19166d37813d23fa5a5db13de365f43c79de2bfc62" Oct 04 06:10:39 crc kubenswrapper[4575]: I1004 06:10:39.403209 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:10:39 crc kubenswrapper[4575]: E1004 06:10:39.403438 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:10:40 crc kubenswrapper[4575]: I1004 06:10:40.082198 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sqpc2_must-gather-l5gpf_fa422c76-44f8-4120-ba4b-facc21cfbdc5/gather/0.log" Oct 04 06:10:49 crc kubenswrapper[4575]: I1004 06:10:49.399333 4575 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sqpc2/must-gather-l5gpf"] Oct 04 06:10:49 crc kubenswrapper[4575]: I1004 06:10:49.400093 4575 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" podUID="fa422c76-44f8-4120-ba4b-facc21cfbdc5" containerName="copy" containerID="cri-o://1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2" gracePeriod=2 Oct 04 06:10:49 crc kubenswrapper[4575]: I1004 06:10:49.432537 4575 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sqpc2/must-gather-l5gpf"] Oct 04 06:10:49 crc kubenswrapper[4575]: I1004 06:10:49.914474 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sqpc2_must-gather-l5gpf_fa422c76-44f8-4120-ba4b-facc21cfbdc5/copy/0.log" Oct 04 06:10:49 crc kubenswrapper[4575]: I1004 06:10:49.915749 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.048675 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa422c76-44f8-4120-ba4b-facc21cfbdc5-must-gather-output\") pod \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\" (UID: \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\") " Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.048898 4575 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2thw\" (UniqueName: \"kubernetes.io/projected/fa422c76-44f8-4120-ba4b-facc21cfbdc5-kube-api-access-r2thw\") pod \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\" (UID: \"fa422c76-44f8-4120-ba4b-facc21cfbdc5\") " Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.074994 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa422c76-44f8-4120-ba4b-facc21cfbdc5-kube-api-access-r2thw" (OuterVolumeSpecName: "kube-api-access-r2thw") pod "fa422c76-44f8-4120-ba4b-facc21cfbdc5" (UID: "fa422c76-44f8-4120-ba4b-facc21cfbdc5"). InnerVolumeSpecName "kube-api-access-r2thw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.151023 4575 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2thw\" (UniqueName: \"kubernetes.io/projected/fa422c76-44f8-4120-ba4b-facc21cfbdc5-kube-api-access-r2thw\") on node \"crc\" DevicePath \"\"" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.303554 4575 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa422c76-44f8-4120-ba4b-facc21cfbdc5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fa422c76-44f8-4120-ba4b-facc21cfbdc5" (UID: "fa422c76-44f8-4120-ba4b-facc21cfbdc5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.312094 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:10:50 crc kubenswrapper[4575]: E1004 06:10:50.312330 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.356527 4575 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa422c76-44f8-4120-ba4b-facc21cfbdc5-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.549785 4575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sqpc2_must-gather-l5gpf_fa422c76-44f8-4120-ba4b-facc21cfbdc5/copy/0.log" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.550228 4575 generic.go:334] "Generic (PLEG): container finished" podID="fa422c76-44f8-4120-ba4b-facc21cfbdc5" containerID="1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2" exitCode=143 Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.550308 4575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sqpc2/must-gather-l5gpf" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.560745 4575 scope.go:117] "RemoveContainer" containerID="1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.604450 4575 scope.go:117] "RemoveContainer" containerID="4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.675848 4575 scope.go:117] "RemoveContainer" containerID="1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2" Oct 04 06:10:50 crc kubenswrapper[4575]: E1004 06:10:50.676469 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2\": container with ID starting with 1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2 not found: ID does not exist" containerID="1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.676502 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2"} err="failed to get container status \"1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2\": rpc error: code = NotFound desc = could not find container \"1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2\": container with ID starting with 1006b8df0444e281ea25d356955d33f5b68782e8cae02fdc71ce841aa86f90e2 not found: ID does not exist" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.676521 4575 scope.go:117] "RemoveContainer" containerID="4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552" Oct 04 06:10:50 crc kubenswrapper[4575]: E1004 06:10:50.677085 4575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552\": container with ID starting with 4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552 not found: ID does not exist" containerID="4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552" Oct 04 06:10:50 crc kubenswrapper[4575]: I1004 06:10:50.677109 4575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552"} err="failed to get container status \"4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552\": rpc error: code = NotFound desc = could not find container \"4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552\": container with ID starting with 4bbd8b3df1336eb66cff53cae74847f81c72c1ad765b29076c3839fe56932552 not found: ID does not exist" Oct 04 06:10:51 crc kubenswrapper[4575]: I1004 06:10:51.319863 4575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa422c76-44f8-4120-ba4b-facc21cfbdc5" path="/var/lib/kubelet/pods/fa422c76-44f8-4120-ba4b-facc21cfbdc5/volumes" Oct 04 06:11:04 crc kubenswrapper[4575]: I1004 06:11:04.311368 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:11:04 crc kubenswrapper[4575]: E1004 06:11:04.312575 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:11:18 crc kubenswrapper[4575]: I1004 06:11:18.311395 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:11:18 crc kubenswrapper[4575]: E1004 06:11:18.312258 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:11:30 crc kubenswrapper[4575]: I1004 06:11:30.310674 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:11:30 crc kubenswrapper[4575]: E1004 06:11:30.311432 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:11:41 crc kubenswrapper[4575]: I1004 06:11:41.310546 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:11:41 crc kubenswrapper[4575]: E1004 06:11:41.311374 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:11:53 crc kubenswrapper[4575]: I1004 06:11:53.310397 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:11:53 crc kubenswrapper[4575]: E1004 06:11:53.311633 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:12:06 crc kubenswrapper[4575]: I1004 06:12:06.310340 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:12:06 crc kubenswrapper[4575]: E1004 06:12:06.311127 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:12:19 crc kubenswrapper[4575]: I1004 06:12:19.309751 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:12:19 crc kubenswrapper[4575]: E1004 06:12:19.310576 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:12:30 crc kubenswrapper[4575]: I1004 06:12:30.310802 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:12:30 crc kubenswrapper[4575]: E1004 06:12:30.311648 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" Oct 04 06:12:33 crc kubenswrapper[4575]: I1004 06:12:33.025783 4575 scope.go:117] "RemoveContainer" containerID="23788336d4df191196abb84a31ae095a726f72d3c46a33f6c747fb1e8bbbb1df" Oct 04 06:12:44 crc kubenswrapper[4575]: I1004 06:12:44.311353 4575 scope.go:117] "RemoveContainer" containerID="3219c0fbff4d572acb279e767d9c12649afa21a134ed7b4ea541c68886ea5143" Oct 04 06:12:44 crc kubenswrapper[4575]: E1004 06:12:44.313376 4575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pl4jq_openshift-machine-config-operator(df9f335e-3c50-4cbf-ad47-491eb2486de8)\"" pod="openshift-machine-config-operator/machine-config-daemon-pl4jq" podUID="df9f335e-3c50-4cbf-ad47-491eb2486de8" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070135345024450 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070135345017365 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070121233016477 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070121234015450 5ustar corecore